fix spelling in a few more files
This commit is contained in:
@ -83,9 +83,9 @@ void PotentialFileReader::next_dvector(double * list, int n) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
ValueTokenizer PotentialFileReader::next_values(int nparams, const std::string & seperators) {
|
ValueTokenizer PotentialFileReader::next_values(int nparams, const std::string & separators) {
|
||||||
try {
|
try {
|
||||||
return reader->next_values(nparams, seperators);
|
return reader->next_values(nparams, separators);
|
||||||
} catch (FileReaderException & e) {
|
} catch (FileReaderException & e) {
|
||||||
error->one(FLERR, e.what());
|
error->one(FLERR, e.what());
|
||||||
}
|
}
|
||||||
|
|||||||
@ -116,6 +116,6 @@ void TextFileReader::next_dvector(double * list, int n) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
ValueTokenizer TextFileReader::next_values(int nparams, const std::string & seperators) {
|
ValueTokenizer TextFileReader::next_values(int nparams, const std::string & separators) {
|
||||||
return ValueTokenizer(next_line(nparams), seperators);
|
return ValueTokenizer(next_line(nparams), separators);
|
||||||
}
|
}
|
||||||
|
|||||||
@ -368,18 +368,18 @@ std::string utils::trim_comment(const std::string & line) {
|
|||||||
Return number of words
|
Return number of words
|
||||||
------------------------------------------------------------------------- */
|
------------------------------------------------------------------------- */
|
||||||
|
|
||||||
size_t utils::count_words(const std::string & text, const std::string & seperators) {
|
size_t utils::count_words(const std::string & text, const std::string & separators) {
|
||||||
size_t count = 0;
|
size_t count = 0;
|
||||||
size_t start = text.find_first_not_of(seperators);
|
size_t start = text.find_first_not_of(separators);
|
||||||
|
|
||||||
while (start != std::string::npos) {
|
while (start != std::string::npos) {
|
||||||
size_t end = text.find_first_of(seperators, start);
|
size_t end = text.find_first_of(separators, start);
|
||||||
++count;
|
++count;
|
||||||
|
|
||||||
if(end == std::string::npos) {
|
if(end == std::string::npos) {
|
||||||
return count;
|
return count;
|
||||||
} else {
|
} else {
|
||||||
start = text.find_first_not_of(seperators, end + 1);
|
start = text.find_first_not_of(separators, end + 1);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return count;
|
return count;
|
||||||
@ -389,8 +389,8 @@ size_t utils::count_words(const std::string & text, const std::string & seperato
|
|||||||
Trim comment from string and return number of words
|
Trim comment from string and return number of words
|
||||||
------------------------------------------------------------------------- */
|
------------------------------------------------------------------------- */
|
||||||
|
|
||||||
size_t utils::trim_and_count_words(const std::string & text, const std::string & seperators) {
|
size_t utils::trim_and_count_words(const std::string & text, const std::string & separators) {
|
||||||
return utils::count_words(utils::trim_comment(text), seperators);
|
return utils::count_words(utils::trim_comment(text), separators);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* ----------------------------------------------------------------------
|
/* ----------------------------------------------------------------------
|
||||||
|
|||||||
@ -153,18 +153,18 @@ namespace LAMMPS_NS {
|
|||||||
/**
|
/**
|
||||||
* \brief Count words in string
|
* \brief Count words in string
|
||||||
* \param text string that should be searched
|
* \param text string that should be searched
|
||||||
* \param seperators string containing characters that will be treated as whitespace
|
* \param separators string containing characters that will be treated as whitespace
|
||||||
* \return number of words found
|
* \return number of words found
|
||||||
*/
|
*/
|
||||||
size_t count_words(const std::string & text, const std::string & seperators = " \t\r\n\f");
|
size_t count_words(const std::string & text, const std::string & separators = " \t\r\n\f");
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* \brief Count words in a single line, trim anything from '#' onward
|
* \brief Count words in a single line, trim anything from '#' onward
|
||||||
* \param text string that should be trimmed and searched
|
* \param text string that should be trimmed and searched
|
||||||
* \param seperators string containing characters that will be treated as whitespace
|
* \param separators string containing characters that will be treated as whitespace
|
||||||
* \return number of words found
|
* \return number of words found
|
||||||
*/
|
*/
|
||||||
size_t trim_and_count_words(const std::string & text, const std::string & seperators = " \t\r\n\f");
|
size_t trim_and_count_words(const std::string & text, const std::string & separators = " \t\r\n\f");
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* \brief Check if string can be converted to valid integer
|
* \brief Check if string can be converted to valid integer
|
||||||
|
|||||||
@ -38,12 +38,12 @@ TEST(Tokenizer, two_words) {
|
|||||||
ASSERT_EQ(t.count(), 2);
|
ASSERT_EQ(t.count(), 2);
|
||||||
}
|
}
|
||||||
|
|
||||||
TEST(Tokenizer, prefix_seperators) {
|
TEST(Tokenizer, prefix_separators) {
|
||||||
Tokenizer t(" test word", " ");
|
Tokenizer t(" test word", " ");
|
||||||
ASSERT_EQ(t.count(), 2);
|
ASSERT_EQ(t.count(), 2);
|
||||||
}
|
}
|
||||||
|
|
||||||
TEST(Tokenizer, postfix_seperators) {
|
TEST(Tokenizer, postfix_separators) {
|
||||||
Tokenizer t("test word ", " ");
|
Tokenizer t("test word ", " ");
|
||||||
ASSERT_EQ(t.count(), 2);
|
ASSERT_EQ(t.count(), 2);
|
||||||
}
|
}
|
||||||
@ -55,7 +55,7 @@ TEST(Tokenizer, iterate_words) {
|
|||||||
ASSERT_EQ(t.count(), 2);
|
ASSERT_EQ(t.count(), 2);
|
||||||
}
|
}
|
||||||
|
|
||||||
TEST(Tokenizer, default_seperators) {
|
TEST(Tokenizer, default_separators) {
|
||||||
Tokenizer t(" \r\n test \t word \f");
|
Tokenizer t(" \r\n test \t word \f");
|
||||||
ASSERT_THAT(t.next(), Eq("test"));
|
ASSERT_THAT(t.next(), Eq("test"));
|
||||||
ASSERT_THAT(t.next(), Eq("word"));
|
ASSERT_THAT(t.next(), Eq("word"));
|
||||||
|
|||||||
Reference in New Issue
Block a user