Skip to content

Commit

Permalink
Merge pull request redpanda-data#4916 from andrwng/typos
Browse files Browse the repository at this point in the history
fix various typos
  • Loading branch information
dotnwat committed May 25, 2022
2 parents c4fc0a2 + 3c6f893 commit 2949fc3
Show file tree
Hide file tree
Showing 9 changed files with 10 additions and 10 deletions.
2 changes: 1 addition & 1 deletion cmake/oss.cmake.in
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@ set(base_env
PKG_CONFIG_PATH=@PKG_CONFIG_PATH@
# Workaround for https://savannah.gnu.org/support/?110503
# Starting with autoconf 2.70, it is expected that gtkdocize is installed.
# Since we don't build with thirdparty with docs, point gtkdocize elsewhere.
# Since we don't build thirdparty with docs, point gtkdocize elsewhere.
GTKDOCIZE=/bin/true
)
set(configure_env
Expand Down
2 changes: 1 addition & 1 deletion src/v/cluster/topics_frontend.cc
Original file line number Diff line number Diff line change
Expand Up @@ -643,7 +643,7 @@ ss::future<std::error_code> topics_frontend::finish_moving_partition_replicas(
ss::future<result<model::offset>> topics_frontend::stm_linearizable_barrier(
model::timeout_clock::time_point timeout) {
return _stm.invoke_on(controller_stm_shard, [timeout](controller_stm& stm) {
return stm.instert_linerizable_barrier(timeout);
return stm.insert_linerizable_barrier(timeout);
});
}

Expand Down
2 changes: 1 addition & 1 deletion src/v/coproc/reconciliation_backend.cc
Original file line number Diff line number Diff line change
Expand Up @@ -310,7 +310,7 @@ bool reconciliation_backend::stale_create_non_replicable_partition_request(
const std::vector<model::broker_shard>& replicas) {
/// If the source partition exists in the topics table, but not in
/// the cluster::partition_manager, either this request occurred before the
/// source partition has been instered into the partition_manager and a
/// source partition has been inserted into the partition_manager and a
/// retry of creating 'this' partition should occur (until success), OR the
/// source topic has been moved to another broker_shard and this request is
/// therefore stale
Expand Down
4 changes: 2 additions & 2 deletions src/v/kafka/server/handlers/topics/topic_utils.h
Original file line number Diff line number Diff line change
Expand Up @@ -142,7 +142,7 @@ auto to_cluster_type(KafkaApiTypeIter begin, KafkaApiTypeIter end)

/// Generate errors for all the request items that topic names
/// are duplicated within given range,
/// the errors are insterted in the range begginning at out_it
/// the errors are inserted in the range begginning at out_it
// clang-format off
template<typename Iter, typename ErrIter>
CONCEPT(requires TopicRequestItem<typename Iter::value_type> &&
Expand All @@ -165,7 +165,7 @@ Iter validate_range_duplicates(Iter begin, Iter end, ErrIter out_it) {
}

/// Generate NOT_CONTROLLER error for all the request items within given range
/// the errors are insterted in the range begginning at out_it
/// the errors are inserted in the range begginning at out_it
/// This pattern is used in every Admin request of Kafka protocol.
template<typename Iter, typename ErrIter>
CONCEPT(requires TopicRequestItem<typename Iter::value_type>&&
Expand Down
2 changes: 1 addition & 1 deletion src/v/raft/configuration_manager.cc
Original file line number Diff line number Diff line change
Expand Up @@ -71,7 +71,7 @@ ss::future<> configuration_manager::prefix_truncate(model::offset offset) {
"Configuration manager should always have at least one "
"configuration");
/**
* When prefix truncation would remove all the configuration we instert
* When prefix truncation would remove all the configuration we insert
* the last configuration from before requested offset at the offset.
* This way we preserver last know configuration and indexing.
*
Expand Down
2 changes: 1 addition & 1 deletion src/v/raft/replicate_entries_stm.cc
Original file line number Diff line number Diff line change
Expand Up @@ -233,7 +233,7 @@ inline bool replicate_entries_stm::should_skip_follower_request(vnode id) {
if (it->second.last_received_append_entries_reply_timestamp < timeout) {
vlog(
_ctxlog.trace,
"Skipping sending append request to {} - didn'r receive "
"Skipping sending append request to {} - didn't receive "
"follower heartbeat",
id);
return true;
Expand Down
2 changes: 1 addition & 1 deletion src/v/raft/state_machine.cc
Original file line number Diff line number Diff line change
Expand Up @@ -144,7 +144,7 @@ ss::future<> state_machine::write_last_applied(model::offset o) {
return _raft->write_last_applied(o);
}

ss::future<result<model::offset>> state_machine::instert_linerizable_barrier(
ss::future<result<model::offset>> state_machine::insert_linerizable_barrier(
model::timeout_clock::time_point timeout) {
return ss::with_timeout(timeout, _raft->linearizable_barrier())
.handle_exception_type([](const ss::timed_out_error&) {
Expand Down
2 changes: 1 addition & 1 deletion src/v/raft/state_machine.h
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,7 @@ class state_machine {
* details see paragraph 6.4 of Raft protocol dissertation.
*/
ss::future<result<model::offset>>
instert_linerizable_barrier(model::timeout_clock::time_point);
insert_linerizable_barrier(model::timeout_clock::time_point);

protected:
void set_next(model::offset offset);
Expand Down
2 changes: 1 addition & 1 deletion src/v/storage/readers_cache.cc
Original file line number Diff line number Diff line change
Expand Up @@ -139,7 +139,7 @@ ss::future<> readers_cache::stop() {
}
/**
* First we close the gate, this will prevent new readers from being
* insterted to the cache
* inserted to the cache
*/
co_await _gate.close();
/**
Expand Down

0 comments on commit 2949fc3

Please sign in to comment.