mirror of
https://github.com/bspeice/dtparse
synced 2024-12-22 04:18:09 -05:00
Add notes on codegen, and regenerate
This commit is contained in:
parent
50fe2c01d4
commit
0d3b646749
@ -189,6 +189,10 @@ def test_fuzzy_simple(i, s):
|
|||||||
|
|
||||||
# Here lies all the ugly junk.
|
# Here lies all the ugly junk.
|
||||||
TEST_HEADER = '''
|
TEST_HEADER = '''
|
||||||
|
//! This code has been generated by running the `build_pycompat.py` script
|
||||||
|
//! in the repository root. Please do not edit it, as your edits will be destroyed
|
||||||
|
//! upon re-running code generation.
|
||||||
|
|
||||||
extern crate chrono;
|
extern crate chrono;
|
||||||
|
|
||||||
use chrono::Datelike;
|
use chrono::Datelike;
|
||||||
|
@ -24,6 +24,10 @@ fn test_tokenize{i}() {{
|
|||||||
|
|
||||||
|
|
||||||
TEST_HEADER = '''
|
TEST_HEADER = '''
|
||||||
|
//! This code has been generated by running the `build_pycompat_tokenizer.py` script
|
||||||
|
//! in the repository root. Please do not edit it, as your edits will be destroyed
|
||||||
|
//! upon re-running code generation.
|
||||||
|
|
||||||
use tokenize::Tokenizer;
|
use tokenize::Tokenizer;
|
||||||
|
|
||||||
fn tokenize_assert(test_str: &str, comparison: Vec<&str>) {
|
fn tokenize_assert(test_str: &str, comparison: Vec<&str>) {
|
||||||
|
@ -1,4 +1,8 @@
|
|||||||
|
|
||||||
|
//! This code has been generated by running the `build_pycompat.py` script
|
||||||
|
//! in the repository root. Please do not edit it, as your edits will be destroyed
|
||||||
|
//! upon re-running code generation.
|
||||||
|
|
||||||
extern crate chrono;
|
extern crate chrono;
|
||||||
|
|
||||||
use chrono::Datelike;
|
use chrono::Datelike;
|
||||||
|
@ -1,4 +1,8 @@
|
|||||||
|
|
||||||
|
//! This code has been generated by running the `build_pycompat_tokenizer.py` script
|
||||||
|
//! in the repository root. Please do not edit it, as your edits will be destroyed
|
||||||
|
//! upon re-running code generation.
|
||||||
|
|
||||||
use tokenize::Tokenizer;
|
use tokenize::Tokenizer;
|
||||||
|
|
||||||
fn tokenize_assert(test_str: &str, comparison: Vec<&str>) {
|
fn tokenize_assert(test_str: &str, comparison: Vec<&str>) {
|
||||||
|
Loading…
Reference in New Issue
Block a user