mirror of
https://github.com/n08i40k/schedule-parser-rusted.git
synced 2025-12-06 17:57:47 +03:00
Compare commits
15 Commits
release/v1
...
d23092a32a
| Author | SHA1 | Date | |
|---|---|---|---|
|
d23092a32a
|
|||
|
01bfa38969
|
|||
|
851ec9225f
|
|||
|
8de1891724
|
|||
|
4cf6df379e
|
|||
|
ba8b164b6a
|
|||
|
ff9d7d6c3a
|
|||
|
9090716f87
|
|||
|
ee992f1b55
|
|||
|
7f71fb1616
|
|||
|
234055eaeb
|
|||
|
fceffb900d
|
|||
|
49ce0005dc
|
|||
|
4c738085f2
|
|||
|
20602eb863
|
3
.idea/schedule-parser-rusted.iml
generated
3
.idea/schedule-parser-rusted.iml
generated
@@ -4,9 +4,10 @@
|
|||||||
<content url="file://$MODULE_DIR$">
|
<content url="file://$MODULE_DIR$">
|
||||||
<sourceFolder url="file://$MODULE_DIR$/lib/schedule_parser/src" isTestSource="false" />
|
<sourceFolder url="file://$MODULE_DIR$/lib/schedule_parser/src" isTestSource="false" />
|
||||||
<sourceFolder url="file://$MODULE_DIR$/src" isTestSource="false" />
|
<sourceFolder url="file://$MODULE_DIR$/src" isTestSource="false" />
|
||||||
<sourceFolder url="file://$MODULE_DIR$/benches" isTestSource="true" />
|
|
||||||
<sourceFolder url="file://$MODULE_DIR$/actix-macros/src" isTestSource="false" />
|
<sourceFolder url="file://$MODULE_DIR$/actix-macros/src" isTestSource="false" />
|
||||||
<sourceFolder url="file://$MODULE_DIR$/actix-test/src" isTestSource="false" />
|
<sourceFolder url="file://$MODULE_DIR$/actix-test/src" isTestSource="false" />
|
||||||
|
<sourceFolder url="file://$MODULE_DIR$/schedule-parser/benches" isTestSource="true" />
|
||||||
|
<sourceFolder url="file://$MODULE_DIR$/schedule-parser/src" isTestSource="false" />
|
||||||
<excludeFolder url="file://$MODULE_DIR$/actix-macros/target" />
|
<excludeFolder url="file://$MODULE_DIR$/actix-macros/target" />
|
||||||
<excludeFolder url="file://$MODULE_DIR$/actix-test/target" />
|
<excludeFolder url="file://$MODULE_DIR$/actix-test/target" />
|
||||||
<excludeFolder url="file://$MODULE_DIR$/target" />
|
<excludeFolder url="file://$MODULE_DIR$/target" />
|
||||||
|
|||||||
174
Cargo.lock
generated
174
Cargo.lock
generated
@@ -696,6 +696,15 @@ version = "0.10.0"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "1cb3c4a0d3776f7535c32793be81d6d5fec0d48ac70955d9834e643aa249a52f"
|
checksum = "1cb3c4a0d3776f7535c32793be81d6d5fec0d48ac70955d9834e643aa249a52f"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "convert_case"
|
||||||
|
version = "0.7.1"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "bb402b8d4c85569410425650ce3eddc7d698ed96d39a73f941b08fb63082f1e7"
|
||||||
|
dependencies = [
|
||||||
|
"unicode-segmentation",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "cookie"
|
name = "cookie"
|
||||||
version = "0.16.2"
|
version = "0.16.2"
|
||||||
@@ -753,25 +762,22 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "criterion"
|
name = "criterion"
|
||||||
version = "0.5.1"
|
version = "0.6.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "f2b12d017a929603d80db1831cd3a24082f8137ce19c69e6447f54f5fc8d692f"
|
checksum = "3bf7af66b0989381bd0be551bd7cc91912a655a58c6918420c9527b1fd8b4679"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"anes",
|
"anes",
|
||||||
"cast",
|
"cast",
|
||||||
"ciborium",
|
"ciborium",
|
||||||
"clap",
|
"clap",
|
||||||
"criterion-plot",
|
"criterion-plot",
|
||||||
"is-terminal",
|
"itertools 0.13.0",
|
||||||
"itertools",
|
|
||||||
"num-traits",
|
"num-traits",
|
||||||
"once_cell",
|
|
||||||
"oorandom",
|
"oorandom",
|
||||||
"plotters",
|
"plotters",
|
||||||
"rayon",
|
"rayon",
|
||||||
"regex",
|
"regex",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_derive",
|
|
||||||
"serde_json",
|
"serde_json",
|
||||||
"tinytemplate",
|
"tinytemplate",
|
||||||
"walkdir",
|
"walkdir",
|
||||||
@@ -784,7 +790,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
|||||||
checksum = "6b50826342786a51a89e2da3a28f1c32b06e387201bc2d19791f622c673706b1"
|
checksum = "6b50826342786a51a89e2da3a28f1c32b06e387201bc2d19791f622c673706b1"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"cast",
|
"cast",
|
||||||
"itertools",
|
"itertools 0.10.5",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
@@ -918,6 +924,7 @@ version = "2.0.1"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "bda628edc44c4bb645fbe0f758797143e4e07926f7ebf4e9bdfbd3d2ce621df3"
|
checksum = "bda628edc44c4bb645fbe0f758797143e4e07926f7ebf4e9bdfbd3d2ce621df3"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
|
"convert_case",
|
||||||
"proc-macro2",
|
"proc-macro2",
|
||||||
"quote",
|
"quote",
|
||||||
"syn 2.0.100",
|
"syn 2.0.100",
|
||||||
@@ -1251,15 +1258,6 @@ dependencies = [
|
|||||||
"slab",
|
"slab",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "fuzzy-matcher"
|
|
||||||
version = "0.3.7"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "54614a3312934d066701a80f20f15fa3b56d67ac7722b39eea5b4c9dd1d66c94"
|
|
||||||
dependencies = [
|
|
||||||
"thread_local",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "gcc"
|
name = "gcc"
|
||||||
version = "0.3.55"
|
version = "0.3.55"
|
||||||
@@ -1404,12 +1402,6 @@ version = "0.5.0"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "2304e00983f87ffb38b55b444b5e3b60a884b5d30c0fca7d82fe33449bbe55ea"
|
checksum = "2304e00983f87ffb38b55b444b5e3b60a884b5d30c0fca7d82fe33449bbe55ea"
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "hermit-abi"
|
|
||||||
version = "0.5.0"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "fbd780fe5cc30f81464441920d82ac8740e2e46b29a6fad543ddd075229ce37e"
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "hex"
|
name = "hex"
|
||||||
version = "0.4.3"
|
version = "0.4.3"
|
||||||
@@ -1848,17 +1840,6 @@ version = "2.11.0"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "469fb0b9cefa57e3ef31275ee7cacb78f2fdca44e4765491884a2b119d4eb130"
|
checksum = "469fb0b9cefa57e3ef31275ee7cacb78f2fdca44e4765491884a2b119d4eb130"
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "is-terminal"
|
|
||||||
version = "0.4.16"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "e04d7f318608d35d4b61ddd75cbdaee86b023ebe2bd5a66ee0915f0bf93095a9"
|
|
||||||
dependencies = [
|
|
||||||
"hermit-abi",
|
|
||||||
"libc",
|
|
||||||
"windows-sys 0.59.0",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "is_terminal_polyfill"
|
name = "is_terminal_polyfill"
|
||||||
version = "1.70.1"
|
version = "1.70.1"
|
||||||
@@ -1874,6 +1855,15 @@ dependencies = [
|
|||||||
"either",
|
"either",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "itertools"
|
||||||
|
version = "0.13.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "413ee7dfc52ee1a4949ceeb7dbc8a33f2d6c088194d9f922fb8318faf1f01186"
|
||||||
|
dependencies = [
|
||||||
|
"either",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "itoa"
|
name = "itoa"
|
||||||
version = "1.0.15"
|
version = "1.0.15"
|
||||||
@@ -2396,7 +2386,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
|||||||
checksum = "8a56d757972c98b346a9b766e3f02746cde6dd1cd1d1d563472929fdd74bec4d"
|
checksum = "8a56d757972c98b346a9b766e3f02746cde6dd1cd1d1d563472929fdd74bec4d"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"anyhow",
|
"anyhow",
|
||||||
"itertools",
|
"itertools 0.10.5",
|
||||||
"proc-macro2",
|
"proc-macro2",
|
||||||
"quote",
|
"quote",
|
||||||
"syn 2.0.100",
|
"syn 2.0.100",
|
||||||
@@ -2441,7 +2431,7 @@ dependencies = [
|
|||||||
"rustc-hash",
|
"rustc-hash",
|
||||||
"rustls",
|
"rustls",
|
||||||
"socket2",
|
"socket2",
|
||||||
"thiserror 2.0.12",
|
"thiserror",
|
||||||
"tokio",
|
"tokio",
|
||||||
"tracing",
|
"tracing",
|
||||||
"web-time",
|
"web-time",
|
||||||
@@ -2461,7 +2451,7 @@ dependencies = [
|
|||||||
"rustls",
|
"rustls",
|
||||||
"rustls-pki-types",
|
"rustls-pki-types",
|
||||||
"slab",
|
"slab",
|
||||||
"thiserror 2.0.12",
|
"thiserror",
|
||||||
"tinyvec",
|
"tinyvec",
|
||||||
"tracing",
|
"tracing",
|
||||||
"web-time",
|
"web-time",
|
||||||
@@ -2874,17 +2864,31 @@ dependencies = [
|
|||||||
"windows-sys 0.59.0",
|
"windows-sys 0.59.0",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "schedule-parser"
|
||||||
|
version = "0.1.0"
|
||||||
|
dependencies = [
|
||||||
|
"calamine",
|
||||||
|
"chrono",
|
||||||
|
"criterion",
|
||||||
|
"derive_more",
|
||||||
|
"regex",
|
||||||
|
"sentry",
|
||||||
|
"serde",
|
||||||
|
"serde_repr",
|
||||||
|
"strsim",
|
||||||
|
"utoipa",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "schedule-parser-rusted"
|
name = "schedule-parser-rusted"
|
||||||
version = "1.0.1"
|
version = "1.0.5"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"actix-macros 0.1.0",
|
"actix-macros 0.1.0",
|
||||||
"actix-test",
|
"actix-test",
|
||||||
"actix-web",
|
"actix-web",
|
||||||
"bcrypt",
|
"bcrypt",
|
||||||
"calamine",
|
|
||||||
"chrono",
|
"chrono",
|
||||||
"criterion",
|
|
||||||
"derive_more",
|
"derive_more",
|
||||||
"diesel",
|
"diesel",
|
||||||
"diesel-derive-enum",
|
"diesel-derive-enum",
|
||||||
@@ -2892,19 +2896,17 @@ dependencies = [
|
|||||||
"env_logger",
|
"env_logger",
|
||||||
"firebase-messaging-rs",
|
"firebase-messaging-rs",
|
||||||
"futures-util",
|
"futures-util",
|
||||||
"fuzzy-matcher",
|
|
||||||
"hex",
|
"hex",
|
||||||
"jsonwebtoken",
|
"jsonwebtoken",
|
||||||
"mime",
|
"mime",
|
||||||
"objectid",
|
"objectid",
|
||||||
"rand 0.9.0",
|
"rand 0.9.0",
|
||||||
"regex",
|
|
||||||
"reqwest",
|
"reqwest",
|
||||||
|
"schedule-parser",
|
||||||
"sentry",
|
"sentry",
|
||||||
"sentry-actix",
|
"sentry-actix",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
"serde_repr",
|
|
||||||
"serde_with",
|
"serde_with",
|
||||||
"sha1 0.11.0-pre.5",
|
"sha1 0.11.0-pre.5",
|
||||||
"tokio",
|
"tokio",
|
||||||
@@ -2977,13 +2979,14 @@ checksum = "56e6fa9c48d24d85fb3de5ad847117517440f6beceb7798af16b4a87d616b8d0"
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "sentry"
|
name = "sentry"
|
||||||
version = "0.37.0"
|
version = "0.38.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "255914a8e53822abd946e2ce8baa41d4cded6b8e938913b7f7b9da5b7ab44335"
|
checksum = "a505499b38861edd82b5a688fa06ba4ba5875bb832adeeeba22b7b23fc4bc39a"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"httpdate",
|
"httpdate",
|
||||||
"native-tls",
|
"native-tls",
|
||||||
"reqwest",
|
"reqwest",
|
||||||
|
"sentry-actix",
|
||||||
"sentry-backtrace",
|
"sentry-backtrace",
|
||||||
"sentry-contexts",
|
"sentry-contexts",
|
||||||
"sentry-core",
|
"sentry-core",
|
||||||
@@ -2996,9 +2999,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "sentry-actix"
|
name = "sentry-actix"
|
||||||
version = "0.37.0"
|
version = "0.38.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "a927aed43cce0e9240f7477ac81cdfa2ffb048e0e2b17000eb5976e14f063993"
|
checksum = "39ad8bfdcfbc6e0d0dacaa5728555085ef459fa9226cfc2fe64eefa4b8038b7f"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"actix-http",
|
"actix-http",
|
||||||
"actix-web",
|
"actix-web",
|
||||||
@@ -3009,21 +3012,20 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "sentry-backtrace"
|
name = "sentry-backtrace"
|
||||||
version = "0.37.0"
|
version = "0.38.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "00293cd332a859961f24fd69258f7e92af736feaeb91020cff84dac4188a4302"
|
checksum = "8dace796060e4ad10e3d1405b122ae184a8b2e71dce05ae450e4f81b7686b0d9"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"backtrace",
|
"backtrace",
|
||||||
"once_cell",
|
|
||||||
"regex",
|
"regex",
|
||||||
"sentry-core",
|
"sentry-core",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "sentry-contexts"
|
name = "sentry-contexts"
|
||||||
version = "0.37.0"
|
version = "0.38.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "961990f9caa76476c481de130ada05614cd7f5aa70fb57c2142f0e09ad3fb2aa"
|
checksum = "87bd9e6b51ffe2bc7188ebe36cb67557cb95749c08a3f81f33e8c9b135e0d1bc"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"hostname 0.4.1",
|
"hostname 0.4.1",
|
||||||
"libc",
|
"libc",
|
||||||
@@ -3035,12 +3037,11 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "sentry-core"
|
name = "sentry-core"
|
||||||
version = "0.37.0"
|
version = "0.38.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "1a6409d845707d82415c800290a5d63be5e3df3c2e417b0997c60531dfbd35ef"
|
checksum = "7426d4beec270cfdbb50f85f0bb2ce176ea57eed0b11741182a163055a558187"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"once_cell",
|
"rand 0.9.0",
|
||||||
"rand 0.8.5",
|
|
||||||
"sentry-types",
|
"sentry-types",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
@@ -3048,20 +3049,19 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "sentry-debug-images"
|
name = "sentry-debug-images"
|
||||||
version = "0.37.0"
|
version = "0.38.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "71ab5df4f3b64760508edfe0ba4290feab5acbbda7566a79d72673065888e5cc"
|
checksum = "9df15c066c04f34c4dfd496a8e76590106b93283f72ef1a47d8fb24d88493424"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"findshlibs",
|
"findshlibs",
|
||||||
"once_cell",
|
|
||||||
"sentry-core",
|
"sentry-core",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "sentry-panic"
|
name = "sentry-panic"
|
||||||
version = "0.37.0"
|
version = "0.38.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "609b1a12340495ce17baeec9e08ff8ed423c337c1a84dffae36a178c783623f3"
|
checksum = "c92beed69b776a162b6d269bef1eaa3e614090b6df45a88d9b239c4fdbffdfba"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"sentry-backtrace",
|
"sentry-backtrace",
|
||||||
"sentry-core",
|
"sentry-core",
|
||||||
@@ -3069,9 +3069,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "sentry-tracing"
|
name = "sentry-tracing"
|
||||||
version = "0.37.0"
|
version = "0.38.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "49f4e86402d5c50239dc7d8fd3f6d5e048221d5fcb4e026d8d50ab57fe4644cb"
|
checksum = "55c323492795de90824f3198562e33dd74ae3bc852fbb13c0cabec54a1cf73cd"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"sentry-backtrace",
|
"sentry-backtrace",
|
||||||
"sentry-core",
|
"sentry-core",
|
||||||
@@ -3081,16 +3081,16 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "sentry-types"
|
name = "sentry-types"
|
||||||
version = "0.37.0"
|
version = "0.38.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "3d3f117b8755dbede8260952de2aeb029e20f432e72634e8969af34324591631"
|
checksum = "04b6c9287202294685cb1f749b944dbbce8160b81a1061ecddc073025fed129f"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"debugid",
|
"debugid",
|
||||||
"hex",
|
"hex",
|
||||||
"rand 0.8.5",
|
"rand 0.9.0",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
"thiserror 1.0.69",
|
"thiserror",
|
||||||
"time 0.3.40",
|
"time 0.3.40",
|
||||||
"url",
|
"url",
|
||||||
"uuid",
|
"uuid",
|
||||||
@@ -3232,7 +3232,7 @@ checksum = "297f631f50729c8c99b84667867963997ec0b50f32b2a7dbcab828ef0541e8bb"
|
|||||||
dependencies = [
|
dependencies = [
|
||||||
"num-bigint",
|
"num-bigint",
|
||||||
"num-traits",
|
"num-traits",
|
||||||
"thiserror 2.0.12",
|
"thiserror",
|
||||||
"time 0.3.40",
|
"time 0.3.40",
|
||||||
]
|
]
|
||||||
|
|
||||||
@@ -3355,33 +3355,13 @@ dependencies = [
|
|||||||
"windows-sys 0.59.0",
|
"windows-sys 0.59.0",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "thiserror"
|
|
||||||
version = "1.0.69"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "b6aaf5339b578ea85b50e080feb250a3e8ae8cfcdff9a461c9ec2904bc923f52"
|
|
||||||
dependencies = [
|
|
||||||
"thiserror-impl 1.0.69",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "thiserror"
|
name = "thiserror"
|
||||||
version = "2.0.12"
|
version = "2.0.12"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "567b8a2dae586314f7be2a752ec7474332959c6460e02bde30d702a66d488708"
|
checksum = "567b8a2dae586314f7be2a752ec7474332959c6460e02bde30d702a66d488708"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"thiserror-impl 2.0.12",
|
"thiserror-impl",
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "thiserror-impl"
|
|
||||||
version = "1.0.69"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "4fee6c4efc90059e10f81e6d42c60a18f76588c3d74cb83a0b242a2b6c7504c1"
|
|
||||||
dependencies = [
|
|
||||||
"proc-macro2",
|
|
||||||
"quote",
|
|
||||||
"syn 2.0.100",
|
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
@@ -3395,16 +3375,6 @@ dependencies = [
|
|||||||
"syn 2.0.100",
|
"syn 2.0.100",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "thread_local"
|
|
||||||
version = "1.1.8"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "8b9ef9bad013ada3808854ceac7b46812a6465ba368859a37e2100283d2d719c"
|
|
||||||
dependencies = [
|
|
||||||
"cfg-if",
|
|
||||||
"once_cell",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "time"
|
name = "time"
|
||||||
version = "0.1.45"
|
version = "0.1.45"
|
||||||
@@ -3722,6 +3692,12 @@ version = "1.0.18"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "5a5f39404a5da50712a4c1eecf25e90dd62b613502b7e925fd4e4d19b5c96512"
|
checksum = "5a5f39404a5da50712a4c1eecf25e90dd62b613502b7e925fd4e4d19b5c96512"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "unicode-segmentation"
|
||||||
|
version = "1.12.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "f6ccf251212114b54433ec949fd6a7841275f9ada20dddd2f29e9ceea4501493"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "unicode-xid"
|
name = "unicode-xid"
|
||||||
version = "0.2.6"
|
version = "0.2.6"
|
||||||
@@ -4367,7 +4343,7 @@ dependencies = [
|
|||||||
"flate2",
|
"flate2",
|
||||||
"indexmap 2.8.0",
|
"indexmap 2.8.0",
|
||||||
"memchr",
|
"memchr",
|
||||||
"thiserror 2.0.12",
|
"thiserror",
|
||||||
"zopfli",
|
"zopfli",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|||||||
20
Cargo.toml
20
Cargo.toml
@@ -1,9 +1,9 @@
|
|||||||
[workspace]
|
[workspace]
|
||||||
members = ["actix-macros", "actix-test"]
|
members = ["actix-macros", "actix-test", "schedule-parser"]
|
||||||
|
|
||||||
[package]
|
[package]
|
||||||
name = "schedule-parser-rusted"
|
name = "schedule-parser-rusted"
|
||||||
version = "1.0.1"
|
version = "1.0.5"
|
||||||
edition = "2024"
|
edition = "2024"
|
||||||
publish = false
|
publish = false
|
||||||
|
|
||||||
@@ -13,29 +13,26 @@ debug = true
|
|||||||
[dependencies]
|
[dependencies]
|
||||||
actix-web = "4.10.2"
|
actix-web = "4.10.2"
|
||||||
actix-macros = { path = "actix-macros" }
|
actix-macros = { path = "actix-macros" }
|
||||||
|
schedule-parser = { path = "schedule-parser", features = ["test-utils"] }
|
||||||
bcrypt = "0.17.0"
|
bcrypt = "0.17.0"
|
||||||
calamine = "0.26.1"
|
|
||||||
chrono = { version = "0.4.40", features = ["serde"] }
|
chrono = { version = "0.4.40", features = ["serde"] }
|
||||||
derive_more = "2.0.1"
|
derive_more = { version = "2", features = ["full"] }
|
||||||
diesel = { version = "2.2.8", features = ["postgres"] }
|
diesel = { version = "2.2.8", features = ["postgres"] }
|
||||||
diesel-derive-enum = { git = "https://github.com/Havunen/diesel-derive-enum.git", features = ["postgres"] }
|
diesel-derive-enum = { git = "https://github.com/Havunen/diesel-derive-enum.git", features = ["postgres"] }
|
||||||
dotenvy = "0.15.7"
|
dotenvy = "0.15.7"
|
||||||
env_logger = "0.11.7"
|
env_logger = "0.11.7"
|
||||||
firebase-messaging-rs = { git = "https://github.com/i10416/firebase-messaging-rs.git" }
|
firebase-messaging-rs = { git = "https://github.com/i10416/firebase-messaging-rs.git" }
|
||||||
futures-util = "0.3.31"
|
futures-util = "0.3.31"
|
||||||
fuzzy-matcher = "0.3.7"
|
|
||||||
jsonwebtoken = { version = "9.3.1", features = ["use_pem"] }
|
jsonwebtoken = { version = "9.3.1", features = ["use_pem"] }
|
||||||
hex = "0.4.3"
|
hex = "0.4.3"
|
||||||
mime = "0.3.17"
|
mime = "0.3.17"
|
||||||
objectid = "0.2.0"
|
objectid = "0.2.0"
|
||||||
regex = "1.11.1"
|
|
||||||
reqwest = { version = "0.12.15", features = ["json"] }
|
reqwest = { version = "0.12.15", features = ["json"] }
|
||||||
sentry = "0.37.0"
|
sentry = "0.38"
|
||||||
sentry-actix = "0.37.0"
|
sentry-actix = "0.38"
|
||||||
serde = { version = "1.0.219", features = ["derive"] }
|
serde = { version = "1.0.219", features = ["derive"] }
|
||||||
serde_json = "1.0.140"
|
serde_json = "1.0.140"
|
||||||
serde_with = "3.12.0"
|
serde_with = "3.12.0"
|
||||||
serde_repr = "0.1.20"
|
|
||||||
sha1 = "0.11.0-pre.5"
|
sha1 = "0.11.0-pre.5"
|
||||||
tokio = { version = "1.44.1", features = ["macros", "rt-multi-thread"] }
|
tokio = { version = "1.44.1", features = ["macros", "rt-multi-thread"] }
|
||||||
rand = "0.9.0"
|
rand = "0.9.0"
|
||||||
@@ -46,8 +43,3 @@ uuid = { version = "1.16.0", features = ["v4"] }
|
|||||||
|
|
||||||
[dev-dependencies]
|
[dev-dependencies]
|
||||||
actix-test = { path = "actix-test" }
|
actix-test = { path = "actix-test" }
|
||||||
criterion = "0.5.1"
|
|
||||||
|
|
||||||
[[bench]]
|
|
||||||
name = "parse"
|
|
||||||
harness = false
|
|
||||||
@@ -6,7 +6,7 @@ ARG BINARY_NAME
|
|||||||
WORKDIR /app/
|
WORKDIR /app/
|
||||||
|
|
||||||
RUN apt update && \
|
RUN apt update && \
|
||||||
apt install -y libpq5
|
apt install -y libpq5 ca-certificates openssl
|
||||||
|
|
||||||
COPY ./${BINARY_NAME} /bin/main
|
COPY ./${BINARY_NAME} /bin/main
|
||||||
RUN chmod +x /bin/main
|
RUN chmod +x /bin/main
|
||||||
|
|||||||
25
schedule-parser/Cargo.toml
Normal file
25
schedule-parser/Cargo.toml
Normal file
@@ -0,0 +1,25 @@
|
|||||||
|
[package]
|
||||||
|
name = "schedule-parser"
|
||||||
|
version = "0.1.0"
|
||||||
|
edition = "2024"
|
||||||
|
|
||||||
|
[features]
|
||||||
|
test-utils = []
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
calamine = "0.26"
|
||||||
|
chrono = { version = "0.4", features = ["serde"] }
|
||||||
|
derive_more = { version = "2", features = ["full"] }
|
||||||
|
sentry = "0.38"
|
||||||
|
serde = { version = "1.0.219", features = ["derive"] }
|
||||||
|
serde_repr = "0.1.20"
|
||||||
|
regex = "1.11.1"
|
||||||
|
utoipa = { version = "5", features = ["chrono"] }
|
||||||
|
strsim = "0.11.1"
|
||||||
|
|
||||||
|
[dev-dependencies]
|
||||||
|
criterion = "0.6"
|
||||||
|
|
||||||
|
[[bench]]
|
||||||
|
name = "parse"
|
||||||
|
harness = false
|
||||||
@@ -1,9 +1,9 @@
|
|||||||
use criterion::{Criterion, criterion_group, criterion_main};
|
use criterion::{Criterion, criterion_group, criterion_main};
|
||||||
|
|
||||||
use schedule_parser_rusted::parser::parse_xls;
|
use schedule_parser::parse_xls;
|
||||||
|
|
||||||
pub fn bench_parse_xls(c: &mut Criterion) {
|
pub fn bench_parse_xls(c: &mut Criterion) {
|
||||||
let buffer: Vec<u8> = include_bytes!("../schedule.xls").to_vec();
|
let buffer: Vec<u8> = include_bytes!("../../schedule.xls").to_vec();
|
||||||
|
|
||||||
c.bench_function("parse_xls", |b| b.iter(|| parse_xls(&buffer).unwrap()));
|
c.bench_function("parse_xls", |b| b.iter(|| parse_xls(&buffer).unwrap()));
|
||||||
}
|
}
|
||||||
765
schedule-parser/src/lib.rs
Normal file
765
schedule-parser/src/lib.rs
Normal file
@@ -0,0 +1,765 @@
|
|||||||
|
use crate::LessonParseResult::{Lessons, Street};
|
||||||
|
use crate::schema::LessonType::Break;
|
||||||
|
use crate::schema::internal::{BoundariesCellInfo, DayCellInfo, GroupCellInfo};
|
||||||
|
use crate::schema::{
|
||||||
|
Day, ErrorCell, ErrorCellPos, Lesson, LessonBoundaries, LessonSubGroup, LessonType, ParseError,
|
||||||
|
ParseResult, ScheduleEntry,
|
||||||
|
};
|
||||||
|
use crate::worksheet::WorkSheet;
|
||||||
|
use calamine::{Reader, Xls, open_workbook_from_rs};
|
||||||
|
use chrono::{DateTime, Duration, NaiveDate, NaiveTime, Utc};
|
||||||
|
use regex::Regex;
|
||||||
|
use std::collections::HashMap;
|
||||||
|
use std::io::Cursor;
|
||||||
|
use std::sync::LazyLock;
|
||||||
|
|
||||||
|
mod macros;
|
||||||
|
pub mod schema;
|
||||||
|
mod worksheet;
|
||||||
|
|
||||||
|
/// Obtaining a "skeleton" schedule from the working sheet.
|
||||||
|
fn parse_skeleton(
|
||||||
|
worksheet: &WorkSheet,
|
||||||
|
) -> Result<(Vec<DayCellInfo>, Vec<GroupCellInfo>), ParseError> {
|
||||||
|
let mut groups: Vec<GroupCellInfo> = Vec::new();
|
||||||
|
let mut days: Vec<DayCellInfo> = Vec::new();
|
||||||
|
|
||||||
|
let worksheet_start = worksheet.start().ok_or(ParseError::UnknownWorkSheetRange)?;
|
||||||
|
let worksheet_end = worksheet.end().ok_or(ParseError::UnknownWorkSheetRange)?;
|
||||||
|
|
||||||
|
let mut row = worksheet_start.0;
|
||||||
|
|
||||||
|
while row < worksheet_end.0 {
|
||||||
|
row += 1;
|
||||||
|
|
||||||
|
let day_full_name = or_continue!(worksheet.get_string_from_cell(row, 0));
|
||||||
|
|
||||||
|
// parse groups row when days column will found
|
||||||
|
if groups.is_empty() {
|
||||||
|
// переход на предыдущую строку
|
||||||
|
row -= 1;
|
||||||
|
|
||||||
|
for column in (worksheet_start.1 + 2)..=worksheet_end.1 {
|
||||||
|
groups.push(GroupCellInfo {
|
||||||
|
column,
|
||||||
|
name: or_continue!(worksheet.get_string_from_cell(row, column)),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// возврат на текущую строку
|
||||||
|
row += 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
let (day_name, day_date) = {
|
||||||
|
let space_index = day_full_name.find(' ').unwrap();
|
||||||
|
|
||||||
|
let name = day_full_name[..space_index].to_string();
|
||||||
|
|
||||||
|
let date_slice = &day_full_name[space_index + 1..];
|
||||||
|
let date = or_break!(NaiveDate::parse_from_str(date_slice, "%d.%m.%Y").ok())
|
||||||
|
.and_time(NaiveTime::default())
|
||||||
|
.and_utc();
|
||||||
|
|
||||||
|
(name, date)
|
||||||
|
};
|
||||||
|
|
||||||
|
days.push(DayCellInfo {
|
||||||
|
row,
|
||||||
|
column: 0,
|
||||||
|
name: day_name,
|
||||||
|
date: day_date,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok((days, groups))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// The result of obtaining a lesson from the cell.
|
||||||
|
enum LessonParseResult {
|
||||||
|
/// List of lessons long from one to two.
|
||||||
|
///
|
||||||
|
/// The number of lessons will be equal to one if the couple is the first in the day,
|
||||||
|
/// otherwise the list from the change template and the lesson itself will be returned.
|
||||||
|
Lessons(Vec<Lesson>),
|
||||||
|
|
||||||
|
/// Street on which the Polytechnic Corps is located.
|
||||||
|
Street(String),
|
||||||
|
}
|
||||||
|
|
||||||
|
// noinspection GrazieInspection
|
||||||
|
/// Obtaining a non-standard type of lesson by name.
|
||||||
|
fn guess_lesson_type(text: &String) -> Option<LessonType> {
|
||||||
|
static MAP: LazyLock<HashMap<&str, LessonType>> = LazyLock::new(|| {
|
||||||
|
HashMap::from([
|
||||||
|
("консультация", LessonType::Consultation),
|
||||||
|
("самостоятельная работа", LessonType::IndependentWork),
|
||||||
|
("зачет", LessonType::Exam),
|
||||||
|
("зачет с оценкой", LessonType::ExamWithGrade),
|
||||||
|
("экзамен", LessonType::ExamDefault),
|
||||||
|
("курсовой проект", LessonType::CourseProject),
|
||||||
|
("защита курсового проекта", LessonType::CourseProjectDefense),
|
||||||
|
])
|
||||||
|
});
|
||||||
|
|
||||||
|
let name_lower = text.to_lowercase();
|
||||||
|
|
||||||
|
match MAP
|
||||||
|
.iter()
|
||||||
|
.map(|(text, lesson_type)| (lesson_type, strsim::levenshtein(text, &*name_lower)))
|
||||||
|
.filter(|x| x.1 <= 4)
|
||||||
|
.min_by_key(|(_, score)| *score)
|
||||||
|
{
|
||||||
|
None => None,
|
||||||
|
Some(v) => Some(v.0.clone()),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Getting a pair or street from a cell.
|
||||||
|
fn parse_lesson(
|
||||||
|
worksheet: &WorkSheet,
|
||||||
|
day: &Day,
|
||||||
|
day_boundaries: &Vec<BoundariesCellInfo>,
|
||||||
|
lesson_boundaries: &BoundariesCellInfo,
|
||||||
|
group_column: u32,
|
||||||
|
) -> Result<LessonParseResult, ParseError> {
|
||||||
|
let row = lesson_boundaries.xls_range.0.0;
|
||||||
|
|
||||||
|
let name = {
|
||||||
|
let cell_data = match worksheet.get_string_from_cell(row, group_column) {
|
||||||
|
Some(x) => x,
|
||||||
|
None => return Ok(Lessons(Vec::new())),
|
||||||
|
};
|
||||||
|
|
||||||
|
static OTHER_STREET_RE: LazyLock<Regex> =
|
||||||
|
LazyLock::new(|| Regex::new(r"^[А-Я][а-я]+[,\s]\d+$").unwrap());
|
||||||
|
|
||||||
|
if OTHER_STREET_RE.is_match(&cell_data) {
|
||||||
|
return Ok(Street(cell_data));
|
||||||
|
}
|
||||||
|
|
||||||
|
cell_data
|
||||||
|
};
|
||||||
|
|
||||||
|
let cell_range = worksheet.get_merge_from_start(row, group_column);
|
||||||
|
|
||||||
|
let (default_range, lesson_time) = {
|
||||||
|
let end_time_arr = day_boundaries
|
||||||
|
.iter()
|
||||||
|
.filter(|time| time.xls_range.1.0 == cell_range.1.0)
|
||||||
|
.collect::<Vec<&BoundariesCellInfo>>();
|
||||||
|
|
||||||
|
let end_time =
|
||||||
|
end_time_arr
|
||||||
|
.first()
|
||||||
|
.ok_or(ParseError::LessonTimeNotFound(ErrorCellPos {
|
||||||
|
row,
|
||||||
|
column: group_column,
|
||||||
|
}))?;
|
||||||
|
|
||||||
|
let range: Option<[u8; 2]> = if lesson_boundaries.default_index != None {
|
||||||
|
let default = lesson_boundaries.default_index.unwrap() as u8;
|
||||||
|
Some([default, end_time.default_index.unwrap() as u8])
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
};
|
||||||
|
|
||||||
|
let time = LessonBoundaries {
|
||||||
|
start: lesson_boundaries.time_range.start,
|
||||||
|
end: end_time.time_range.end,
|
||||||
|
};
|
||||||
|
|
||||||
|
Ok((range, time))
|
||||||
|
}?;
|
||||||
|
|
||||||
|
let (name, mut subgroups, lesson_type) = parse_name_and_subgroups(&name)?;
|
||||||
|
|
||||||
|
{
|
||||||
|
let cabinets: Vec<String> = parse_cabinets(
|
||||||
|
worksheet,
|
||||||
|
(cell_range.0.0, cell_range.1.0),
|
||||||
|
group_column + 1,
|
||||||
|
);
|
||||||
|
|
||||||
|
match cabinets.len() {
|
||||||
|
// Если кабинетов нет, но есть подгруппы, назначаем им кабинет "??"
|
||||||
|
0 => {
|
||||||
|
for subgroup in &mut subgroups {
|
||||||
|
subgroup.cabinet = Some("??".to_string());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Назначаем этот кабинет всем подгруппам
|
||||||
|
1 => {
|
||||||
|
for subgroup in &mut subgroups {
|
||||||
|
subgroup.cabinet =
|
||||||
|
Some(cabinets.get(0).or(Some(&String::new())).unwrap().clone())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
len => {
|
||||||
|
// Если количество кабинетов совпадает с количеством подгрупп, назначаем кабинеты по порядку
|
||||||
|
if len == subgroups.len() {
|
||||||
|
for subgroup in &mut subgroups {
|
||||||
|
subgroup.cabinet = Some(
|
||||||
|
cabinets
|
||||||
|
.get((subgroup.number - 1) as usize)
|
||||||
|
.unwrap()
|
||||||
|
.clone(),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
// Если количество кабинетов больше количества подгрупп, делаем ещё одну подгруппу.
|
||||||
|
} else if len > subgroups.len() {
|
||||||
|
for index in 0..subgroups.len() {
|
||||||
|
subgroups[index].cabinet = Some(cabinets[index].clone());
|
||||||
|
}
|
||||||
|
|
||||||
|
while cabinets.len() > subgroups.len() {
|
||||||
|
subgroups.push(LessonSubGroup {
|
||||||
|
number: (subgroups.len() + 1) as u8,
|
||||||
|
cabinet: Some(cabinets[subgroups.len()].clone()),
|
||||||
|
teacher: "Ошибка в расписании".to_string(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
|
let lesson = Lesson {
|
||||||
|
lesson_type: lesson_type.unwrap_or(lesson_boundaries.lesson_type.clone()),
|
||||||
|
default_range,
|
||||||
|
name: Some(name),
|
||||||
|
time: lesson_time,
|
||||||
|
subgroups: Some(subgroups),
|
||||||
|
group: None,
|
||||||
|
};
|
||||||
|
|
||||||
|
let prev_lesson = if day.lessons.is_empty() {
|
||||||
|
return Ok(Lessons(Vec::from([lesson])));
|
||||||
|
} else {
|
||||||
|
&day.lessons[day.lessons.len() - 1]
|
||||||
|
};
|
||||||
|
|
||||||
|
Ok(Lessons(Vec::from([
|
||||||
|
Lesson {
|
||||||
|
lesson_type: Break,
|
||||||
|
default_range: None,
|
||||||
|
name: None,
|
||||||
|
time: LessonBoundaries {
|
||||||
|
start: prev_lesson.time.end,
|
||||||
|
end: lesson.time.start,
|
||||||
|
},
|
||||||
|
subgroups: Some(Vec::new()),
|
||||||
|
group: None,
|
||||||
|
},
|
||||||
|
lesson,
|
||||||
|
])))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Obtaining a list of cabinets to the right of the lesson cell.
|
||||||
|
fn parse_cabinets(worksheet: &WorkSheet, row_range: (u32, u32), column: u32) -> Vec<String> {
|
||||||
|
let mut cabinets: Vec<String> = Vec::new();
|
||||||
|
|
||||||
|
for row in row_range.0..row_range.1 {
|
||||||
|
let raw = or_continue!(worksheet.get_string_from_cell(row, column));
|
||||||
|
|
||||||
|
let clean = raw.replace("\n", " ");
|
||||||
|
let parts: Vec<&str> = clean.split(" ").collect();
|
||||||
|
|
||||||
|
for part in parts {
|
||||||
|
let clean_part = part.to_string().trim().to_string();
|
||||||
|
|
||||||
|
cabinets.push(clean_part);
|
||||||
|
}
|
||||||
|
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
cabinets
|
||||||
|
}
|
||||||
|
|
||||||
|
//noinspection GrazieInspection
|
||||||
|
/// Getting the "pure" name of the lesson and list of teachers from the text of the lesson cell.
|
||||||
|
fn parse_name_and_subgroups(
|
||||||
|
text: &String,
|
||||||
|
) -> Result<(String, Vec<LessonSubGroup>, Option<LessonType>), ParseError> {
|
||||||
|
// Части названия пары:
|
||||||
|
// 1. Само название.
|
||||||
|
// 2. Список преподавателей и подгрупп.
|
||||||
|
// 3. "Модификатор" (чаще всего).
|
||||||
|
//
|
||||||
|
// Регулярное выражение для получения ФИО преподавателей и номеров подгрупп (aka. второй части).
|
||||||
|
// (?:[А-Я][а-я]+\s?(?:[А-Я][\s.]*){2}(?:\(\d\s?[а-я]+\))?(?:, )?)+[\s.]*
|
||||||
|
//
|
||||||
|
// Подробнее:
|
||||||
|
// (?:
|
||||||
|
// [А-Я][а-я]+ - Фамилия.
|
||||||
|
// \s? - Кто знает, будет ли там пробел.
|
||||||
|
// (?:[А-Я][\s.]*){2} - Имя и отчество с учётом случайных пробелов и точек.
|
||||||
|
// (?:
|
||||||
|
// \( - Открытие подгруппы.
|
||||||
|
// \s? - Кто знает, будет ли там пробел.
|
||||||
|
// \d - Номер подгруппы.
|
||||||
|
// \s? - Кто знает, будет ли там пробел.
|
||||||
|
// [а-я\s]+ - Слово "подгруппа" с учётов ошибок.
|
||||||
|
// \) - Закрытие подгруппы.
|
||||||
|
// )? - Явное указание подгруппы может отсутствовать по понятным причинам.
|
||||||
|
// (?:, )? - Разделители между отдельными частями.
|
||||||
|
// )+
|
||||||
|
// [\s.]* - Забираем с собой всякий мусор, что бы не передать его в третью часть.
|
||||||
|
|
||||||
|
static NAMES_REGEX: LazyLock<Regex> = LazyLock::new(|| {
|
||||||
|
Regex::new(
|
||||||
|
r"(?:[А-Я][а-я]+\s?(?:[А-Я][\s.]*){2}(?:\(\s*\d\s*[а-я\s]+\))?(?:[\s,]+)?)+[\s.]*",
|
||||||
|
)
|
||||||
|
.unwrap()
|
||||||
|
});
|
||||||
|
|
||||||
|
// Отчистка
|
||||||
|
static CLEAN_RE: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"[\s\n\t]+").unwrap());
|
||||||
|
|
||||||
|
let text = CLEAN_RE
|
||||||
|
.replace(&text.replace(&[' ', '\t', '\n'], " "), " ")
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
let (lesson_name, mut subgroups, lesson_type) = match NAMES_REGEX.captures(&text) {
|
||||||
|
Some(captures) => {
|
||||||
|
let capture = captures.get(0).unwrap();
|
||||||
|
|
||||||
|
let subgroups: Vec<LessonSubGroup> = {
|
||||||
|
let src = capture.as_str().replace(&[' ', '.'], "");
|
||||||
|
|
||||||
|
src.split(',')
|
||||||
|
.map(|name| {
|
||||||
|
let open_bracket_index = name.find('(');
|
||||||
|
|
||||||
|
let subgroup_number = open_bracket_index.map_or(0, |index| {
|
||||||
|
name[(index + 1)..(index + 2)].parse::<u8>().unwrap()
|
||||||
|
});
|
||||||
|
|
||||||
|
let teacher_name = {
|
||||||
|
let name_end = open_bracket_index.unwrap_or_else(|| name.len());
|
||||||
|
|
||||||
|
// Я ебал. Как же я долго до этого доходил.
|
||||||
|
format!(
|
||||||
|
"{} {}.{}.",
|
||||||
|
name.get(..name_end - 4).unwrap(),
|
||||||
|
name.get(name_end - 4..name_end - 2).unwrap(),
|
||||||
|
name.get(name_end - 2..name_end).unwrap(),
|
||||||
|
)
|
||||||
|
};
|
||||||
|
|
||||||
|
LessonSubGroup {
|
||||||
|
number: subgroup_number,
|
||||||
|
cabinet: None,
|
||||||
|
teacher: teacher_name,
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.collect()
|
||||||
|
};
|
||||||
|
|
||||||
|
let name = text[..capture.start()].trim().to_string();
|
||||||
|
let extra = text[capture.end()..].trim().to_string();
|
||||||
|
|
||||||
|
let lesson_type = if extra.len() > 4 {
|
||||||
|
let result = guess_lesson_type(&extra);
|
||||||
|
|
||||||
|
#[cfg(not(debug_assertions))]
|
||||||
|
if result.is_none() {
|
||||||
|
sentry::capture_message(
|
||||||
|
&*format!("Не удалось угадать тип пары '{}'!", extra),
|
||||||
|
sentry::Level::Warning,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
result
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
};
|
||||||
|
|
||||||
|
(name, subgroups, lesson_type)
|
||||||
|
}
|
||||||
|
None => (text, Vec::new(), None),
|
||||||
|
};
|
||||||
|
|
||||||
|
// фикс, если у кого-то отсутствует индекс подгруппы
|
||||||
|
|
||||||
|
if subgroups.len() == 1 {
|
||||||
|
let index = subgroups[0].number;
|
||||||
|
|
||||||
|
if index == 0 {
|
||||||
|
subgroups[0].number = 1u8;
|
||||||
|
} else {
|
||||||
|
subgroups.push(LessonSubGroup {
|
||||||
|
number: if index == 1 { 2 } else { 1 },
|
||||||
|
cabinet: None,
|
||||||
|
teacher: "Только у другой".to_string(),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
} else if subgroups.len() == 2 {
|
||||||
|
// если индексы отсутствуют у обоих, ставим поочерёдно
|
||||||
|
if subgroups[0].number == 0 && subgroups[1].number == 0 {
|
||||||
|
subgroups[0].number = 1;
|
||||||
|
subgroups[1].number = 2;
|
||||||
|
}
|
||||||
|
// если индекс отсутствует у первого, ставим 2, если у второго индекс 1 и наоборот
|
||||||
|
else if subgroups[0].number == 0 {
|
||||||
|
subgroups[0].number = if subgroups[1].number == 1 { 2 } else { 1 };
|
||||||
|
}
|
||||||
|
// если индекс отсутствует у второго, ставим 2, если у первого индекс 1 и наоборот
|
||||||
|
else if subgroups[1].number == 0 {
|
||||||
|
subgroups[1].number = if subgroups[0].number == 1 { 2 } else { 1 };
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if subgroups.len() == 2 && subgroups[0].number == 2 && subgroups[1].number == 1 {
|
||||||
|
subgroups.reverse()
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok((lesson_name, subgroups, lesson_type))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Getting the start and end of a pair from a cell in the first column of a document.
|
||||||
|
///
|
||||||
|
/// # Arguments
|
||||||
|
///
|
||||||
|
/// * `cell_data`: text in cell.
|
||||||
|
/// * `date`: date of the current day.
|
||||||
|
fn parse_lesson_boundaries_cell(
|
||||||
|
cell_data: &String,
|
||||||
|
date: DateTime<Utc>,
|
||||||
|
) -> Option<LessonBoundaries> {
|
||||||
|
static TIME_RE: LazyLock<Regex> =
|
||||||
|
LazyLock::new(|| Regex::new(r"(\d+\.\d+)-(\d+\.\d+)").unwrap());
|
||||||
|
|
||||||
|
let parse_res = if let Some(captures) = TIME_RE.captures(cell_data) {
|
||||||
|
captures
|
||||||
|
} else {
|
||||||
|
return None;
|
||||||
|
};
|
||||||
|
|
||||||
|
let start_match = parse_res.get(1).unwrap().as_str();
|
||||||
|
let start_parts: Vec<&str> = start_match.split(".").collect();
|
||||||
|
|
||||||
|
let end_match = parse_res.get(2).unwrap().as_str();
|
||||||
|
let end_parts: Vec<&str> = end_match.split(".").collect();
|
||||||
|
|
||||||
|
static GET_TIME: fn(DateTime<Utc>, &Vec<&str>) -> DateTime<Utc> = |date, parts| {
|
||||||
|
date + Duration::hours(parts[0].parse::<i64>().unwrap() - 4)
|
||||||
|
+ Duration::minutes(parts[1].parse::<i64>().unwrap())
|
||||||
|
};
|
||||||
|
|
||||||
|
Some(LessonBoundaries {
|
||||||
|
start: GET_TIME(date.clone(), &start_parts),
|
||||||
|
end: GET_TIME(date, &end_parts),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Parse the column of the document to obtain a list of day's lesson boundaries.
|
||||||
|
///
|
||||||
|
/// # Arguments
|
||||||
|
///
|
||||||
|
/// * `worksheet`: document.
|
||||||
|
/// * `date`: date of the current day.
|
||||||
|
/// * `row_range`: row boundaries of the current day.
|
||||||
|
/// * `column`: column with the required data.
|
||||||
|
fn parse_day_boundaries(
|
||||||
|
worksheet: &WorkSheet,
|
||||||
|
date: DateTime<Utc>,
|
||||||
|
row_range: (u32, u32),
|
||||||
|
column: u32,
|
||||||
|
) -> Result<Vec<BoundariesCellInfo>, ParseError> {
|
||||||
|
let mut day_times: Vec<BoundariesCellInfo> = Vec::new();
|
||||||
|
|
||||||
|
for row in row_range.0..row_range.1 {
|
||||||
|
let time_cell = if let Some(str) = worksheet.get_string_from_cell(row, column) {
|
||||||
|
str
|
||||||
|
} else {
|
||||||
|
continue;
|
||||||
|
};
|
||||||
|
|
||||||
|
let lesson_time = parse_lesson_boundaries_cell(&time_cell, date.clone()).ok_or(
|
||||||
|
ParseError::LessonBoundaries(ErrorCell::new(row, column, time_cell.clone())),
|
||||||
|
)?;
|
||||||
|
|
||||||
|
// type
|
||||||
|
let lesson_type = if time_cell.contains("пара") {
|
||||||
|
LessonType::Default
|
||||||
|
} else {
|
||||||
|
LessonType::Additional
|
||||||
|
};
|
||||||
|
|
||||||
|
// lesson index
|
||||||
|
let default_index = if lesson_type == LessonType::Default {
|
||||||
|
Some(
|
||||||
|
time_cell
|
||||||
|
.chars()
|
||||||
|
.next()
|
||||||
|
.unwrap()
|
||||||
|
.to_string()
|
||||||
|
.parse::<u32>()
|
||||||
|
.unwrap(),
|
||||||
|
)
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
};
|
||||||
|
|
||||||
|
day_times.push(BoundariesCellInfo {
|
||||||
|
time_range: lesson_time,
|
||||||
|
lesson_type,
|
||||||
|
default_index,
|
||||||
|
xls_range: worksheet.get_merge_from_start(row, column),
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(day_times)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Parse the column of the document to obtain a list of week's lesson boundaries.
|
||||||
|
///
|
||||||
|
/// # Arguments
|
||||||
|
///
|
||||||
|
/// * `worksheet`: document.
|
||||||
|
/// * `week_markup`: markup of the current week.
|
||||||
|
fn parse_week_boundaries(
|
||||||
|
worksheet: &WorkSheet,
|
||||||
|
week_markup: &Vec<DayCellInfo>,
|
||||||
|
) -> Result<Vec<Vec<BoundariesCellInfo>>, ParseError> {
|
||||||
|
let mut result: Vec<Vec<BoundariesCellInfo>> = Vec::new();
|
||||||
|
|
||||||
|
let worksheet_end_row = worksheet.end().unwrap().0;
|
||||||
|
let lesson_time_column = week_markup[0].column + 1;
|
||||||
|
|
||||||
|
for day_index in 0..week_markup.len() {
|
||||||
|
let day_markup = &week_markup[day_index];
|
||||||
|
|
||||||
|
// Если текущий день не последнему, то индекс строки следующего дня.
|
||||||
|
// Если текущий день - последний, то индекс последней строки документа.
|
||||||
|
let end_row = if day_index != week_markup.len() - 1 {
|
||||||
|
week_markup[day_index + 1].row
|
||||||
|
} else {
|
||||||
|
worksheet_end_row
|
||||||
|
};
|
||||||
|
|
||||||
|
let day_boundaries = parse_day_boundaries(
|
||||||
|
&worksheet,
|
||||||
|
day_markup.date.clone(),
|
||||||
|
(day_markup.row, end_row),
|
||||||
|
lesson_time_column,
|
||||||
|
)?;
|
||||||
|
|
||||||
|
result.push(day_boundaries);
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(result)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Conversion of the list of couples of groups in the list of lessons of teachers.
|
||||||
|
fn convert_groups_to_teachers(
|
||||||
|
groups: &HashMap<String, ScheduleEntry>,
|
||||||
|
) -> HashMap<String, ScheduleEntry> {
|
||||||
|
let mut teachers: HashMap<String, ScheduleEntry> = HashMap::new();
|
||||||
|
|
||||||
|
let empty_days: Vec<Day> = groups
|
||||||
|
.values()
|
||||||
|
.next()
|
||||||
|
.unwrap()
|
||||||
|
.days
|
||||||
|
.iter()
|
||||||
|
.map(|day| Day {
|
||||||
|
name: day.name.clone(),
|
||||||
|
street: day.street.clone(),
|
||||||
|
date: day.date.clone(),
|
||||||
|
lessons: vec![],
|
||||||
|
})
|
||||||
|
.collect();
|
||||||
|
|
||||||
|
for group in groups.values() {
|
||||||
|
for (index, day) in group.days.iter().enumerate() {
|
||||||
|
for group_lesson in &day.lessons {
|
||||||
|
if group_lesson.lesson_type == Break {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if group_lesson.subgroups.is_none() {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
let subgroups = group_lesson.subgroups.as_ref().unwrap();
|
||||||
|
|
||||||
|
for subgroup in subgroups {
|
||||||
|
if subgroup.teacher == "Ошибка в расписании" {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if !teachers.contains_key(&subgroup.teacher) {
|
||||||
|
teachers.insert(
|
||||||
|
subgroup.teacher.clone(),
|
||||||
|
ScheduleEntry {
|
||||||
|
name: subgroup.teacher.clone(),
|
||||||
|
days: empty_days.to_vec(),
|
||||||
|
},
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
let teacher_day = teachers
|
||||||
|
.get_mut(&subgroup.teacher)
|
||||||
|
.unwrap()
|
||||||
|
.days
|
||||||
|
.get_mut(index)
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
teacher_day.lessons.push({
|
||||||
|
let mut lesson = group_lesson.clone();
|
||||||
|
lesson.group = Some(group.name.clone());
|
||||||
|
|
||||||
|
lesson
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
teachers.iter_mut().for_each(|(_, teacher)| {
|
||||||
|
teacher.days.iter_mut().for_each(|day| {
|
||||||
|
day.lessons.sort_by(|a, b| {
|
||||||
|
a.default_range.as_ref().unwrap()[1].cmp(&b.default_range.as_ref().unwrap()[1])
|
||||||
|
})
|
||||||
|
})
|
||||||
|
});
|
||||||
|
|
||||||
|
teachers
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Reading XLS Document from the buffer and converting it into the schedule ready to use.
|
||||||
|
///
|
||||||
|
/// # Arguments
|
||||||
|
///
|
||||||
|
/// * `buffer`: XLS data containing schedule.
|
||||||
|
///
|
||||||
|
/// returns: Result<ParseResult, ParseError>
|
||||||
|
///
|
||||||
|
/// # Examples
|
||||||
|
///
|
||||||
|
/// ```
|
||||||
|
/// use schedule_parser::parse_xls;
|
||||||
|
///
|
||||||
|
/// let result = parse_xls(&include_bytes!("../../schedule.xls").to_vec());
|
||||||
|
///
|
||||||
|
/// assert!(result.is_ok(), "{}", result.err().unwrap());
|
||||||
|
///
|
||||||
|
/// assert_ne!(result.as_ref().unwrap().groups.len(), 0);
|
||||||
|
/// assert_ne!(result.as_ref().unwrap().teachers.len(), 0);
|
||||||
|
/// ```
|
||||||
|
pub fn parse_xls(buffer: &Vec<u8>) -> Result<ParseResult, ParseError> {
|
||||||
|
let cursor = Cursor::new(&buffer);
|
||||||
|
let mut workbook: Xls<_> =
|
||||||
|
open_workbook_from_rs(cursor).map_err(|e| ParseError::BadXLS(std::sync::Arc::new(e)))?;
|
||||||
|
|
||||||
|
let worksheet = {
|
||||||
|
let (worksheet_name, worksheet) = workbook
|
||||||
|
.worksheets()
|
||||||
|
.first()
|
||||||
|
.ok_or(ParseError::NoWorkSheets)?
|
||||||
|
.clone();
|
||||||
|
|
||||||
|
let worksheet_merges = workbook
|
||||||
|
.worksheet_merge_cells(&*worksheet_name)
|
||||||
|
.ok_or(ParseError::NoWorkSheets)?;
|
||||||
|
|
||||||
|
WorkSheet {
|
||||||
|
data: worksheet,
|
||||||
|
merges: worksheet_merges,
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
let (week_markup, groups_markup) = parse_skeleton(&worksheet)?;
|
||||||
|
let week_boundaries = parse_week_boundaries(&worksheet, &week_markup)?;
|
||||||
|
|
||||||
|
let mut groups: HashMap<String, ScheduleEntry> = HashMap::new();
|
||||||
|
|
||||||
|
for group_markup in groups_markup {
|
||||||
|
let mut group = ScheduleEntry {
|
||||||
|
name: group_markup.name,
|
||||||
|
days: Vec::new(),
|
||||||
|
};
|
||||||
|
|
||||||
|
for day_index in 0..(&week_markup).len() {
|
||||||
|
let day_markup = &week_markup[day_index];
|
||||||
|
|
||||||
|
let mut day = Day {
|
||||||
|
name: day_markup.name.clone(),
|
||||||
|
street: None,
|
||||||
|
date: day_markup.date,
|
||||||
|
lessons: Vec::new(),
|
||||||
|
};
|
||||||
|
|
||||||
|
let day_boundaries = &week_boundaries[day_index];
|
||||||
|
|
||||||
|
for lesson_boundaries in day_boundaries {
|
||||||
|
match &mut parse_lesson(
|
||||||
|
&worksheet,
|
||||||
|
&day,
|
||||||
|
&day_boundaries,
|
||||||
|
&lesson_boundaries,
|
||||||
|
group_markup.column,
|
||||||
|
)? {
|
||||||
|
Lessons(lesson) => day.lessons.append(lesson),
|
||||||
|
Street(street) => day.street = Some(street.to_owned()),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
group.days.push(day);
|
||||||
|
}
|
||||||
|
|
||||||
|
groups.insert(group.name.clone(), group);
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(ParseResult {
|
||||||
|
teachers: convert_groups_to_teachers(&groups),
|
||||||
|
groups,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(any(test, feature = "test-utils"))]
|
||||||
|
pub mod test_utils {
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
pub fn test_result() -> Result<ParseResult, ParseError> {
|
||||||
|
parse_xls(&include_bytes!("../../schedule.xls").to_vec())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
pub mod tests {
|
||||||
|
#[test]
|
||||||
|
fn read() {
|
||||||
|
let result = super::test_utils::test_result();
|
||||||
|
|
||||||
|
assert!(result.is_ok(), "{}", result.err().unwrap());
|
||||||
|
|
||||||
|
assert_ne!(result.as_ref().unwrap().groups.len(), 0);
|
||||||
|
assert_ne!(result.as_ref().unwrap().teachers.len(), 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_split_lesson() {
|
||||||
|
let result = super::test_utils::test_result();
|
||||||
|
assert!(result.is_ok(), "{}", result.err().unwrap());
|
||||||
|
|
||||||
|
let result = result.unwrap();
|
||||||
|
assert!(result.groups.contains_key("ИС-214/23"));
|
||||||
|
|
||||||
|
let group = result.groups.get("ИС-214/23").unwrap();
|
||||||
|
|
||||||
|
let thursday = group.days.get(3).unwrap();
|
||||||
|
assert_eq!(thursday.lessons.len(), 1);
|
||||||
|
|
||||||
|
let lesson = &thursday.lessons[0];
|
||||||
|
assert_eq!(lesson.default_range.unwrap()[1], 3);
|
||||||
|
assert!(lesson.subgroups.is_some());
|
||||||
|
|
||||||
|
let subgroups = lesson.subgroups.as_ref().unwrap();
|
||||||
|
assert_eq!(subgroups.len(), 2);
|
||||||
|
assert_eq!(subgroups[0].cabinet, Some("44".to_string()));
|
||||||
|
assert_eq!(subgroups[1].cabinet, Some("43".to_string()));
|
||||||
|
}
|
||||||
|
}
|
||||||
25
schedule-parser/src/macros.rs
Normal file
25
schedule-parser/src/macros.rs
Normal file
@@ -0,0 +1,25 @@
|
|||||||
|
#[macro_export]
|
||||||
|
macro_rules! or_continue {
|
||||||
|
( $e:expr ) => {
|
||||||
|
{
|
||||||
|
if let Some(x) = $e {
|
||||||
|
x
|
||||||
|
} else {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[macro_export]
|
||||||
|
macro_rules! or_break {
|
||||||
|
( $e:expr ) => {
|
||||||
|
{
|
||||||
|
if let Some(x) = $e {
|
||||||
|
x
|
||||||
|
} else {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -6,9 +6,53 @@ use std::collections::HashMap;
|
|||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use utoipa::ToSchema;
|
use utoipa::ToSchema;
|
||||||
|
|
||||||
|
pub(crate) mod internal {
|
||||||
|
use crate::schema::{LessonBoundaries, LessonType};
|
||||||
|
use chrono::{DateTime, Utc};
|
||||||
|
|
||||||
|
/// Data cell storing the group name.
|
||||||
|
pub struct GroupCellInfo {
|
||||||
|
/// Column index.
|
||||||
|
pub column: u32,
|
||||||
|
|
||||||
|
/// Text in the cell.
|
||||||
|
pub name: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Data cell storing the line.
|
||||||
|
pub struct DayCellInfo {
|
||||||
|
/// Line index.
|
||||||
|
pub row: u32,
|
||||||
|
|
||||||
|
/// Column index.
|
||||||
|
pub column: u32,
|
||||||
|
|
||||||
|
/// Day name.
|
||||||
|
pub name: String,
|
||||||
|
|
||||||
|
/// Date of the day.
|
||||||
|
pub date: DateTime<Utc>,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Data on the time of lessons from the second column of the schedule.
|
||||||
|
pub struct BoundariesCellInfo {
|
||||||
|
/// Temporary segment of the lesson.
|
||||||
|
pub time_range: LessonBoundaries,
|
||||||
|
|
||||||
|
/// Type of lesson.
|
||||||
|
pub lesson_type: LessonType,
|
||||||
|
|
||||||
|
/// The lesson index.
|
||||||
|
pub default_index: Option<u32>,
|
||||||
|
|
||||||
|
/// The frame of the cell.
|
||||||
|
pub xls_range: ((u32, u32), (u32, u32)),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// The beginning and end of the lesson.
|
/// The beginning and end of the lesson.
|
||||||
#[derive(Clone, Hash, Debug, Serialize, Deserialize, ToSchema)]
|
#[derive(Clone, Hash, Debug, Serialize, Deserialize, ToSchema)]
|
||||||
pub struct LessonTime {
|
pub struct LessonBoundaries {
|
||||||
/// The beginning of a lesson.
|
/// The beginning of a lesson.
|
||||||
pub start: DateTime<Utc>,
|
pub start: DateTime<Utc>,
|
||||||
|
|
||||||
@@ -44,6 +88,12 @@ pub enum LessonType {
|
|||||||
|
|
||||||
/// Экзамен.
|
/// Экзамен.
|
||||||
ExamDefault,
|
ExamDefault,
|
||||||
|
|
||||||
|
/// Курсовой проект.
|
||||||
|
CourseProject,
|
||||||
|
|
||||||
|
/// Защита курсового проекта.
|
||||||
|
CourseProjectDefense,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Hash, Debug, Serialize, Deserialize, ToSchema)]
|
#[derive(Clone, Hash, Debug, Serialize, Deserialize, ToSchema)]
|
||||||
@@ -72,7 +122,7 @@ pub struct Lesson {
|
|||||||
pub name: Option<String>,
|
pub name: Option<String>,
|
||||||
|
|
||||||
/// The beginning and end.
|
/// The beginning and end.
|
||||||
pub time: LessonTime,
|
pub time: LessonBoundaries,
|
||||||
|
|
||||||
/// List of subgroups.
|
/// List of subgroups.
|
||||||
#[serde(rename = "subGroups")]
|
#[serde(rename = "subGroups")]
|
||||||
@@ -153,9 +203,9 @@ pub enum ParseError {
|
|||||||
#[display("There is no data on work sheet boundaries.")]
|
#[display("There is no data on work sheet boundaries.")]
|
||||||
UnknownWorkSheetRange,
|
UnknownWorkSheetRange,
|
||||||
|
|
||||||
/// Failed to read the beginning and end of the lesson from the line
|
/// Failed to read the beginning and end of the lesson from the cell
|
||||||
#[display("Failed to read lesson start and end times from {_0}.")]
|
#[display("Failed to read lesson start and end from {_0}.")]
|
||||||
GlobalTime(ErrorCell),
|
LessonBoundaries(ErrorCell),
|
||||||
|
|
||||||
/// Not found the beginning and the end corresponding to the lesson.
|
/// Not found the beginning and the end corresponding to the lesson.
|
||||||
#[display("No start and end times matching the lesson (at {_0}) was found.")]
|
#[display("No start and end times matching the lesson (at {_0}) was found.")]
|
||||||
@@ -173,7 +223,7 @@ impl Serialize for ParseError {
|
|||||||
ParseError::UnknownWorkSheetRange => {
|
ParseError::UnknownWorkSheetRange => {
|
||||||
serializer.serialize_str("UNKNOWN_WORK_SHEET_RANGE")
|
serializer.serialize_str("UNKNOWN_WORK_SHEET_RANGE")
|
||||||
}
|
}
|
||||||
ParseError::GlobalTime(_) => serializer.serialize_str("GLOBAL_TIME"),
|
ParseError::LessonBoundaries(_) => serializer.serialize_str("GLOBAL_TIME"),
|
||||||
ParseError::LessonTimeNotFound(_) => serializer.serialize_str("LESSON_TIME_NOT_FOUND"),
|
ParseError::LessonTimeNotFound(_) => serializer.serialize_str("LESSON_TIME_NOT_FOUND"),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
58
schedule-parser/src/worksheet.rs
Normal file
58
schedule-parser/src/worksheet.rs
Normal file
@@ -0,0 +1,58 @@
|
|||||||
|
use std::ops::Deref;
|
||||||
|
use std::sync::LazyLock;
|
||||||
|
use regex::Regex;
|
||||||
|
|
||||||
|
/// XLS WorkSheet data.
|
||||||
|
pub struct WorkSheet {
|
||||||
|
pub data: calamine::Range<calamine::Data>,
|
||||||
|
pub merges: Vec<calamine::Dimensions>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Deref for WorkSheet {
|
||||||
|
type Target = calamine::Range<calamine::Data>;
|
||||||
|
|
||||||
|
fn deref(&self) -> &Self::Target {
|
||||||
|
&self.data
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl WorkSheet {
|
||||||
|
/// Getting a line from the required cell.
|
||||||
|
pub fn get_string_from_cell(&self, row: u32, col: u32) -> Option<String> {
|
||||||
|
let cell_data = if let Some(data) = self.get((row as usize, col as usize)) {
|
||||||
|
data.to_string()
|
||||||
|
} else {
|
||||||
|
return None;
|
||||||
|
};
|
||||||
|
|
||||||
|
if cell_data.trim().is_empty() {
|
||||||
|
return None;
|
||||||
|
}
|
||||||
|
|
||||||
|
static NL_RE: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"[\n\r]+").unwrap());
|
||||||
|
static SP_RE: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"\s+").unwrap());
|
||||||
|
|
||||||
|
let trimmed_data = SP_RE
|
||||||
|
.replace_all(&NL_RE.replace_all(&cell_data, " "), " ")
|
||||||
|
.trim()
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
if trimmed_data.is_empty() {
|
||||||
|
None
|
||||||
|
} else {
|
||||||
|
Some(trimmed_data)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Obtaining the boundaries of the cell along its upper left coordinate.
|
||||||
|
pub fn get_merge_from_start(&self, row: u32, column: u32) -> ((u32, u32), (u32, u32)) {
|
||||||
|
match self
|
||||||
|
.merges
|
||||||
|
.iter()
|
||||||
|
.find(|merge| merge.start.0 == row && merge.start.1 == column)
|
||||||
|
{
|
||||||
|
Some(merge) => (merge.start, (merge.end.0 + 1, merge.end.1 + 1)),
|
||||||
|
None => ((row, column), (row + 1, column + 1)),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
BIN
schedule.xls
BIN
schedule.xls
Binary file not shown.
@@ -1,4 +1,4 @@
|
|||||||
use crate::parser::schema::ParseResult;
|
use schedule_parser::schema::ParseResult;
|
||||||
use crate::utility::hasher::DigestHasher;
|
use crate::utility::hasher::DigestHasher;
|
||||||
use crate::xls_downloader::basic_impl::BasicXlsDownloader;
|
use crate::xls_downloader::basic_impl::BasicXlsDownloader;
|
||||||
use actix_web::web;
|
use actix_web::web;
|
||||||
|
|||||||
@@ -1 +0,0 @@
|
|||||||
pub mod parser;
|
|
||||||
@@ -13,7 +13,6 @@ mod app_state;
|
|||||||
|
|
||||||
mod database;
|
mod database;
|
||||||
|
|
||||||
mod parser;
|
|
||||||
mod xls_downloader;
|
mod xls_downloader;
|
||||||
|
|
||||||
mod extractors;
|
mod extractors;
|
||||||
@@ -112,8 +111,6 @@ fn main() -> io::Result<()> {
|
|||||||
},
|
},
|
||||||
));
|
));
|
||||||
|
|
||||||
unsafe { std::env::set_var("RUST_BACKTRACE", "1") };
|
|
||||||
|
|
||||||
dotenv().unwrap();
|
dotenv().unwrap();
|
||||||
|
|
||||||
env_logger::init();
|
env_logger::init();
|
||||||
|
|||||||
@@ -1,743 +0,0 @@
|
|||||||
use crate::parser::LessonParseResult::{Lessons, Street};
|
|
||||||
use crate::parser::schema::LessonType::Break;
|
|
||||||
use crate::parser::schema::{
|
|
||||||
Day, ErrorCell, ErrorCellPos, Lesson, LessonSubGroup, LessonTime, LessonType, ParseError,
|
|
||||||
ParseResult, ScheduleEntry,
|
|
||||||
};
|
|
||||||
use calamine::{Reader, Xls, open_workbook_from_rs};
|
|
||||||
use chrono::{DateTime, Duration, NaiveDateTime, Utc};
|
|
||||||
use fuzzy_matcher::FuzzyMatcher;
|
|
||||||
use fuzzy_matcher::skim::SkimMatcherV2;
|
|
||||||
use regex::Regex;
|
|
||||||
use std::collections::HashMap;
|
|
||||||
use std::io::Cursor;
|
|
||||||
use std::sync::LazyLock;
|
|
||||||
|
|
||||||
pub mod schema;
|
|
||||||
|
|
||||||
/// Data cell storing the line.
|
|
||||||
struct InternalId {
|
|
||||||
/// Line index.
|
|
||||||
row: u32,
|
|
||||||
|
|
||||||
/// Column index.
|
|
||||||
column: u32,
|
|
||||||
|
|
||||||
/// Text in the cell.
|
|
||||||
name: String,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Data on the time of lessons from the second column of the schedule.
|
|
||||||
struct InternalTime {
|
|
||||||
/// Temporary segment of the lesson.
|
|
||||||
time_range: LessonTime,
|
|
||||||
|
|
||||||
/// Type of lesson.
|
|
||||||
lesson_type: LessonType,
|
|
||||||
|
|
||||||
/// The lesson index.
|
|
||||||
default_index: Option<u32>,
|
|
||||||
|
|
||||||
/// The frame of the cell.
|
|
||||||
xls_range: ((u32, u32), (u32, u32)),
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Working sheet type alias.
|
|
||||||
type WorkSheet = calamine::Range<calamine::Data>;
|
|
||||||
|
|
||||||
/// Getting a line from the required cell.
|
|
||||||
fn get_string_from_cell(worksheet: &WorkSheet, row: u32, col: u32) -> Option<String> {
|
|
||||||
let cell_data = if let Some(data) = worksheet.get((row as usize, col as usize)) {
|
|
||||||
data.to_string()
|
|
||||||
} else {
|
|
||||||
return None;
|
|
||||||
};
|
|
||||||
|
|
||||||
if cell_data.trim().is_empty() {
|
|
||||||
return None;
|
|
||||||
}
|
|
||||||
|
|
||||||
static NL_RE: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"[\n\r]+").unwrap());
|
|
||||||
static SP_RE: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"\s+").unwrap());
|
|
||||||
|
|
||||||
let trimmed_data = SP_RE
|
|
||||||
.replace_all(&NL_RE.replace_all(&cell_data, " "), " ")
|
|
||||||
.trim()
|
|
||||||
.to_string();
|
|
||||||
|
|
||||||
if trimmed_data.is_empty() {
|
|
||||||
None
|
|
||||||
} else {
|
|
||||||
Some(trimmed_data)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Obtaining the boundaries of the cell along its upper left coordinate.
|
|
||||||
fn get_merge_from_start(worksheet: &WorkSheet, row: u32, column: u32) -> ((u32, u32), (u32, u32)) {
|
|
||||||
let worksheet_end = worksheet.end().unwrap();
|
|
||||||
|
|
||||||
let row_end: u32 = {
|
|
||||||
let mut r: u32 = 0;
|
|
||||||
|
|
||||||
for _r in (row + 1)..worksheet_end.0 {
|
|
||||||
r = _r;
|
|
||||||
|
|
||||||
if let Some(_) = worksheet.get((_r as usize, column as usize)) {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
r
|
|
||||||
};
|
|
||||||
|
|
||||||
let column_end: u32 = {
|
|
||||||
let mut c: u32 = 0;
|
|
||||||
|
|
||||||
for _c in (column + 1)..worksheet_end.1 {
|
|
||||||
c = _c;
|
|
||||||
|
|
||||||
if let Some(_) = worksheet.get((row as usize, _c as usize)) {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
c
|
|
||||||
};
|
|
||||||
|
|
||||||
((row, column), (row_end, column_end))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Obtaining a "skeleton" schedule from the working sheet.
|
|
||||||
fn parse_skeleton(worksheet: &WorkSheet) -> Result<(Vec<InternalId>, Vec<InternalId>), ParseError> {
|
|
||||||
let range = &worksheet;
|
|
||||||
|
|
||||||
let mut is_parsed = false;
|
|
||||||
|
|
||||||
let mut groups: Vec<InternalId> = Vec::new();
|
|
||||||
let mut days: Vec<InternalId> = Vec::new();
|
|
||||||
|
|
||||||
let start = range.start().ok_or(ParseError::UnknownWorkSheetRange)?;
|
|
||||||
let end = range.end().ok_or(ParseError::UnknownWorkSheetRange)?;
|
|
||||||
|
|
||||||
let mut row = start.0;
|
|
||||||
while row < end.0 {
|
|
||||||
row += 1;
|
|
||||||
|
|
||||||
let day_name_opt = get_string_from_cell(&worksheet, row, 0);
|
|
||||||
if day_name_opt.is_none() {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
let day_name = day_name_opt.unwrap();
|
|
||||||
|
|
||||||
if !is_parsed {
|
|
||||||
is_parsed = true;
|
|
||||||
|
|
||||||
row -= 1;
|
|
||||||
|
|
||||||
for column in (start.1 + 2)..=end.1 {
|
|
||||||
let group_name = get_string_from_cell(&worksheet, row, column);
|
|
||||||
if group_name.is_none() {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
groups.push(InternalId {
|
|
||||||
row,
|
|
||||||
column,
|
|
||||||
name: group_name.unwrap(),
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
row += 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
days.push(InternalId {
|
|
||||||
row,
|
|
||||||
column: 0,
|
|
||||||
name: day_name.clone(),
|
|
||||||
});
|
|
||||||
|
|
||||||
if days.len() > 2 && day_name.starts_with("Суббота") {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok((days, groups))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// The result of obtaining a lesson from the cell.
|
|
||||||
enum LessonParseResult {
|
|
||||||
/// List of lessons long from one to two.
|
|
||||||
///
|
|
||||||
/// The number of lessons will be equal to one if the couple is the first in the day,
|
|
||||||
/// otherwise the list from the change template and the lesson itself will be returned.
|
|
||||||
Lessons(Vec<Lesson>),
|
|
||||||
|
|
||||||
/// Street on which the Polytechnic Corps is located.
|
|
||||||
Street(String),
|
|
||||||
}
|
|
||||||
|
|
||||||
trait StringInnerSlice {
|
|
||||||
/// Obtaining a line from the line on the initial and final index.
|
|
||||||
fn inner_slice(&self, from: usize, to: usize) -> Self;
|
|
||||||
}
|
|
||||||
|
|
||||||
impl StringInnerSlice for String {
|
|
||||||
fn inner_slice(&self, from: usize, to: usize) -> Self {
|
|
||||||
self.chars()
|
|
||||||
.take(from)
|
|
||||||
.chain(self.chars().skip(to))
|
|
||||||
.collect()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// noinspection GrazieInspection
|
|
||||||
/// Obtaining a non-standard type of lesson by name.
|
|
||||||
fn guess_lesson_type(name: &String) -> Option<(String, LessonType)> {
|
|
||||||
let map: HashMap<String, LessonType> = HashMap::from([
|
|
||||||
("(консультация)".to_string(), LessonType::Consultation),
|
|
||||||
(
|
|
||||||
"самостоятельная работа".to_string(),
|
|
||||||
LessonType::IndependentWork,
|
|
||||||
),
|
|
||||||
("зачет".to_string(), LessonType::Exam),
|
|
||||||
("зачет с оценкой".to_string(), LessonType::ExamWithGrade),
|
|
||||||
("экзамен".to_string(), LessonType::ExamDefault),
|
|
||||||
]);
|
|
||||||
|
|
||||||
let matcher = SkimMatcherV2::default();
|
|
||||||
let name_lower = name.to_lowercase();
|
|
||||||
|
|
||||||
type SearchResult<'a> = (&'a LessonType, i64, Vec<usize>);
|
|
||||||
|
|
||||||
let mut search_results: Vec<SearchResult> = map
|
|
||||||
.iter()
|
|
||||||
.map(|entry| -> SearchResult {
|
|
||||||
if let Some((score, indices)) = matcher.fuzzy_indices(&*name_lower, entry.0) {
|
|
||||||
return (entry.1, score, indices);
|
|
||||||
}
|
|
||||||
|
|
||||||
(entry.1, 0, Vec::new())
|
|
||||||
})
|
|
||||||
.collect();
|
|
||||||
search_results.sort_by(|a, b| b.1.cmp(&a.1));
|
|
||||||
|
|
||||||
let guessed_type = search_results.first().unwrap();
|
|
||||||
|
|
||||||
if guessed_type.1 > 80 {
|
|
||||||
Some((
|
|
||||||
name.inner_slice(guessed_type.2[0], guessed_type.2[guessed_type.2.len() - 1]),
|
|
||||||
guessed_type.0.clone(),
|
|
||||||
))
|
|
||||||
} else {
|
|
||||||
None
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Getting a pair or street from a cell.
|
|
||||||
fn parse_lesson(
|
|
||||||
worksheet: &WorkSheet,
|
|
||||||
day: &mut Day,
|
|
||||||
day_times: &Vec<InternalTime>,
|
|
||||||
time: &InternalTime,
|
|
||||||
column: u32,
|
|
||||||
) -> Result<LessonParseResult, ParseError> {
|
|
||||||
let row = time.xls_range.0.0;
|
|
||||||
|
|
||||||
let (name, lesson_type) = {
|
|
||||||
let raw_name_opt = get_string_from_cell(&worksheet, row, column);
|
|
||||||
if raw_name_opt.is_none() {
|
|
||||||
return Ok(Lessons(Vec::new()));
|
|
||||||
}
|
|
||||||
|
|
||||||
let raw_name = raw_name_opt.unwrap();
|
|
||||||
|
|
||||||
static OTHER_STREET_RE: LazyLock<Regex> =
|
|
||||||
LazyLock::new(|| Regex::new(r"^[А-Я][а-я]+,?\s?[0-9]+$").unwrap());
|
|
||||||
|
|
||||||
if OTHER_STREET_RE.is_match(&raw_name) {
|
|
||||||
return Ok(Street(raw_name));
|
|
||||||
}
|
|
||||||
|
|
||||||
if let Some(guess) = guess_lesson_type(&raw_name) {
|
|
||||||
guess
|
|
||||||
} else {
|
|
||||||
(raw_name, time.lesson_type.clone())
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let (default_range, lesson_time) = || -> Result<(Option<[u8; 2]>, LessonTime), ParseError> {
|
|
||||||
// check if multi-lesson
|
|
||||||
let cell_range = get_merge_from_start(worksheet, row, column);
|
|
||||||
|
|
||||||
let end_time_arr = day_times
|
|
||||||
.iter()
|
|
||||||
.filter(|time| time.xls_range.1.0 == cell_range.1.0)
|
|
||||||
.collect::<Vec<&InternalTime>>();
|
|
||||||
|
|
||||||
let end_time = end_time_arr
|
|
||||||
.first()
|
|
||||||
.ok_or(ParseError::LessonTimeNotFound(ErrorCellPos { row, column }))?;
|
|
||||||
|
|
||||||
let range: Option<[u8; 2]> = if time.default_index != None {
|
|
||||||
let default = time.default_index.unwrap() as u8;
|
|
||||||
Some([default, end_time.default_index.unwrap() as u8])
|
|
||||||
} else {
|
|
||||||
None
|
|
||||||
};
|
|
||||||
|
|
||||||
let time = LessonTime {
|
|
||||||
start: time.time_range.start,
|
|
||||||
end: end_time.time_range.end,
|
|
||||||
};
|
|
||||||
|
|
||||||
Ok((range, time))
|
|
||||||
}()?;
|
|
||||||
|
|
||||||
let (name, mut subgroups) = parse_name_and_subgroups(&name)?;
|
|
||||||
|
|
||||||
{
|
|
||||||
let cabinets: Vec<String> = parse_cabinets(worksheet, row, column + 1);
|
|
||||||
|
|
||||||
// Если количество кабинетов равно 1, назначаем этот кабинет всем подгруппам
|
|
||||||
if cabinets.len() == 1 {
|
|
||||||
for subgroup in &mut subgroups {
|
|
||||||
subgroup.cabinet = Some(cabinets.get(0).or(Some(&String::new())).unwrap().clone())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// Если количество кабинетов совпадает с количеством подгрупп, назначаем кабинеты по порядку
|
|
||||||
else if cabinets.len() == subgroups.len() {
|
|
||||||
for subgroup in &mut subgroups {
|
|
||||||
subgroup.cabinet = Some(
|
|
||||||
cabinets
|
|
||||||
.get((subgroup.number - 1) as usize)
|
|
||||||
.unwrap()
|
|
||||||
.clone(),
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// Если количество кабинетов больше количества подгрупп, делаем ещё одну подгруппу.
|
|
||||||
else if cabinets.len() > subgroups.len() {
|
|
||||||
for index in 0..subgroups.len() {
|
|
||||||
subgroups[index].cabinet = Some(cabinets[index].clone());
|
|
||||||
}
|
|
||||||
|
|
||||||
while cabinets.len() > subgroups.len() {
|
|
||||||
subgroups.push(LessonSubGroup {
|
|
||||||
number: (subgroups.len() + 1) as u8,
|
|
||||||
cabinet: Some(cabinets[subgroups.len()].clone()),
|
|
||||||
teacher: "Ошибка в расписании".to_string(),
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// Если кабинетов нет, но есть подгруппы, назначаем им значение "??"
|
|
||||||
else {
|
|
||||||
for subgroup in &mut subgroups {
|
|
||||||
subgroup.cabinet = Some("??".to_string());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
cabinets
|
|
||||||
};
|
|
||||||
|
|
||||||
let lesson = Lesson {
|
|
||||||
lesson_type,
|
|
||||||
default_range,
|
|
||||||
name: Some(name),
|
|
||||||
time: lesson_time,
|
|
||||||
subgroups: Some(subgroups),
|
|
||||||
group: None,
|
|
||||||
};
|
|
||||||
|
|
||||||
let prev_lesson = if day.lessons.len() == 0 {
|
|
||||||
return Ok(Lessons(Vec::from([lesson])));
|
|
||||||
} else {
|
|
||||||
&day.lessons[day.lessons.len() - 1]
|
|
||||||
};
|
|
||||||
|
|
||||||
Ok(Lessons(Vec::from([
|
|
||||||
Lesson {
|
|
||||||
lesson_type: Break,
|
|
||||||
default_range: None,
|
|
||||||
name: None,
|
|
||||||
time: LessonTime {
|
|
||||||
start: prev_lesson.time.end,
|
|
||||||
end: lesson.time.start,
|
|
||||||
},
|
|
||||||
subgroups: Some(Vec::new()),
|
|
||||||
group: None,
|
|
||||||
},
|
|
||||||
lesson,
|
|
||||||
])))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Obtaining a list of cabinets to the right of the lesson cell.
|
|
||||||
fn parse_cabinets(worksheet: &WorkSheet, row: u32, column: u32) -> Vec<String> {
|
|
||||||
let mut cabinets: Vec<String> = Vec::new();
|
|
||||||
|
|
||||||
if let Some(raw) = get_string_from_cell(&worksheet, row, column) {
|
|
||||||
let clean = raw.replace("\n", " ");
|
|
||||||
let parts: Vec<&str> = clean.split(" ").collect();
|
|
||||||
|
|
||||||
for part in parts {
|
|
||||||
let clean_part = part.to_string().trim().to_string();
|
|
||||||
|
|
||||||
cabinets.push(clean_part);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
cabinets
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Getting the "pure" name of the lesson and list of teachers from the text of the lesson cell.
|
|
||||||
fn parse_name_and_subgroups(name: &String) -> Result<(String, Vec<LessonSubGroup>), ParseError> {
|
|
||||||
static LESSON_RE: LazyLock<Regex> =
|
|
||||||
LazyLock::new(|| Regex::new(r"(?:[А-Я][а-я]+[А-Я]{2}(?:\([0-9][а-я]+\))?)+$").unwrap());
|
|
||||||
static TEACHER_RE: LazyLock<Regex> =
|
|
||||||
LazyLock::new(|| Regex::new(r"([А-Я][а-я]+)([А-Я])([А-Я])(?:\(([0-9])[а-я]+\))?").unwrap());
|
|
||||||
static CLEAN_RE: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"[\s.,]+").unwrap());
|
|
||||||
static END_CLEAN_RE: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"[.\s]+$").unwrap());
|
|
||||||
|
|
||||||
let (teachers, lesson_name) = {
|
|
||||||
let clean_name = CLEAN_RE.replace_all(&name, "").to_string();
|
|
||||||
|
|
||||||
if let Some(captures) = LESSON_RE.captures(&clean_name) {
|
|
||||||
let capture = captures.get(0).unwrap();
|
|
||||||
let capture_str = capture.as_str().to_string();
|
|
||||||
let capture_name: String = capture_str.chars().take(5).collect();
|
|
||||||
|
|
||||||
(
|
|
||||||
END_CLEAN_RE.replace(&capture_str, "").to_string(),
|
|
||||||
END_CLEAN_RE
|
|
||||||
.replace(&name[0..name.find(&*capture_name).unwrap()], "")
|
|
||||||
.to_string(),
|
|
||||||
)
|
|
||||||
} else {
|
|
||||||
return Ok((END_CLEAN_RE.replace(&name, "").to_string(), Vec::new()));
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let mut subgroups: Vec<LessonSubGroup> = Vec::new();
|
|
||||||
|
|
||||||
let teacher_it = TEACHER_RE.captures_iter(&teachers);
|
|
||||||
|
|
||||||
for captures in teacher_it {
|
|
||||||
subgroups.push(LessonSubGroup {
|
|
||||||
number: match captures.get(4) {
|
|
||||||
Some(capture) => capture.as_str().to_string().parse::<u8>().unwrap(),
|
|
||||||
None => 0,
|
|
||||||
},
|
|
||||||
cabinet: None,
|
|
||||||
teacher: format!(
|
|
||||||
"{} {}.{}.",
|
|
||||||
captures.get(1).unwrap().as_str().to_string(),
|
|
||||||
captures.get(2).unwrap().as_str().to_string(),
|
|
||||||
captures.get(3).unwrap().as_str().to_string()
|
|
||||||
),
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
// фикс, если у кого-то отсутствует индекс подгруппы
|
|
||||||
|
|
||||||
if subgroups.len() == 1 {
|
|
||||||
let index = subgroups[0].number;
|
|
||||||
|
|
||||||
if index == 0 {
|
|
||||||
subgroups[0].number = 1u8;
|
|
||||||
} else {
|
|
||||||
subgroups.push(LessonSubGroup {
|
|
||||||
number: if index == 1 { 2 } else { 1 },
|
|
||||||
cabinet: None,
|
|
||||||
teacher: "Только у другой".to_string(),
|
|
||||||
});
|
|
||||||
}
|
|
||||||
} else if subgroups.len() == 2 {
|
|
||||||
// если индексы отсутствуют у обоих, ставим поочерёдно
|
|
||||||
if subgroups[0].number == 0 && subgroups[1].number == 0 {
|
|
||||||
subgroups[0].number = 1;
|
|
||||||
subgroups[1].number = 2;
|
|
||||||
}
|
|
||||||
// если индекс отсутствует у первого, ставим 2, если у второго индекс 1 и наоборот
|
|
||||||
else if subgroups[0].number == 0 {
|
|
||||||
subgroups[0].number = if subgroups[1].number == 1 { 2 } else { 1 };
|
|
||||||
}
|
|
||||||
// если индекс отсутствует у второго, ставим 2, если у первого индекс 1 и наоборот
|
|
||||||
else if subgroups[1].number == 0 {
|
|
||||||
subgroups[1].number = if subgroups[0].number == 1 { 2 } else { 1 };
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if subgroups.len() == 2 && subgroups[0].number == 2 && subgroups[1].number == 1 {
|
|
||||||
subgroups.reverse()
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok((lesson_name, subgroups))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Conversion of the list of couples of groups in the list of lessons of teachers.
|
|
||||||
fn convert_groups_to_teachers(
|
|
||||||
groups: &HashMap<String, ScheduleEntry>,
|
|
||||||
) -> HashMap<String, ScheduleEntry> {
|
|
||||||
let mut teachers: HashMap<String, ScheduleEntry> = HashMap::new();
|
|
||||||
|
|
||||||
let empty_days: Vec<Day> = groups
|
|
||||||
.values()
|
|
||||||
.next()
|
|
||||||
.unwrap()
|
|
||||||
.days
|
|
||||||
.iter()
|
|
||||||
.map(|day| Day {
|
|
||||||
name: day.name.clone(),
|
|
||||||
street: day.street.clone(),
|
|
||||||
date: day.date.clone(),
|
|
||||||
lessons: vec![],
|
|
||||||
})
|
|
||||||
.collect();
|
|
||||||
|
|
||||||
for group in groups.values() {
|
|
||||||
for (index, day) in group.days.iter().enumerate() {
|
|
||||||
for group_lesson in &day.lessons {
|
|
||||||
if group_lesson.lesson_type == Break {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
if group_lesson.subgroups.is_none() {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
let subgroups = group_lesson.subgroups.as_ref().unwrap();
|
|
||||||
|
|
||||||
for subgroup in subgroups {
|
|
||||||
if subgroup.teacher == "Ошибка в расписании" {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
if !teachers.contains_key(&subgroup.teacher) {
|
|
||||||
teachers.insert(
|
|
||||||
subgroup.teacher.clone(),
|
|
||||||
ScheduleEntry {
|
|
||||||
name: subgroup.teacher.clone(),
|
|
||||||
days: empty_days.to_vec(),
|
|
||||||
},
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
let teacher_day = teachers
|
|
||||||
.get_mut(&subgroup.teacher)
|
|
||||||
.unwrap()
|
|
||||||
.days
|
|
||||||
.get_mut(index)
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
teacher_day.lessons.push({
|
|
||||||
let mut lesson = group_lesson.clone();
|
|
||||||
lesson.group = Some(group.name.clone());
|
|
||||||
|
|
||||||
lesson
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
teachers.iter_mut().for_each(|(_, teacher)| {
|
|
||||||
teacher.days.iter_mut().for_each(|day| {
|
|
||||||
day.lessons.sort_by(|a, b| {
|
|
||||||
a.default_range.as_ref().unwrap()[1].cmp(&b.default_range.as_ref().unwrap()[1])
|
|
||||||
})
|
|
||||||
})
|
|
||||||
});
|
|
||||||
|
|
||||||
teachers
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Reading XLS Document from the buffer and converting it into the schedule ready to use.
|
|
||||||
///
|
|
||||||
/// # Arguments
|
|
||||||
///
|
|
||||||
/// * `buffer`: XLS data containing schedule.
|
|
||||||
///
|
|
||||||
/// returns: Result<ParseResult, ParseError>
|
|
||||||
///
|
|
||||||
/// # Examples
|
|
||||||
///
|
|
||||||
/// ```
|
|
||||||
/// use schedule_parser_rusted::parser::parse_xls;
|
|
||||||
///
|
|
||||||
/// let result = parse_xls(&include_bytes!("../../schedule.xls").to_vec());
|
|
||||||
///
|
|
||||||
/// assert!(result.is_ok());
|
|
||||||
///
|
|
||||||
/// assert_ne!(result.as_ref().unwrap().groups.len(), 0);
|
|
||||||
/// assert_ne!(result.as_ref().unwrap().teachers.len(), 0);
|
|
||||||
/// ```
|
|
||||||
pub fn parse_xls(buffer: &Vec<u8>) -> Result<ParseResult, ParseError> {
|
|
||||||
let cursor = Cursor::new(&buffer);
|
|
||||||
let mut workbook: Xls<_> =
|
|
||||||
open_workbook_from_rs(cursor).map_err(|e| ParseError::BadXLS(std::sync::Arc::new(e)))?;
|
|
||||||
|
|
||||||
let worksheet: WorkSheet = workbook
|
|
||||||
.worksheets()
|
|
||||||
.first()
|
|
||||||
.ok_or(ParseError::NoWorkSheets)?
|
|
||||||
.1
|
|
||||||
.to_owned();
|
|
||||||
|
|
||||||
let (days_markup, groups_markup) = parse_skeleton(&worksheet)?;
|
|
||||||
|
|
||||||
let mut groups: HashMap<String, ScheduleEntry> = HashMap::new();
|
|
||||||
let mut days_times: Vec<Vec<InternalTime>> = Vec::new();
|
|
||||||
|
|
||||||
let saturday_end_row = worksheet.end().unwrap().0;
|
|
||||||
|
|
||||||
for group_markup in groups_markup {
|
|
||||||
let mut group = ScheduleEntry {
|
|
||||||
name: group_markup.name,
|
|
||||||
days: Vec::new(),
|
|
||||||
};
|
|
||||||
|
|
||||||
for day_index in 0..(&days_markup).len() {
|
|
||||||
let day_markup = &days_markup[day_index];
|
|
||||||
|
|
||||||
let mut day = {
|
|
||||||
let space_index = day_markup.name.find(' ').unwrap();
|
|
||||||
|
|
||||||
let name = day_markup.name[..space_index].to_string();
|
|
||||||
|
|
||||||
let date_raw = day_markup.name[space_index + 1..].to_string();
|
|
||||||
let date_add = format!("{} 00:00:00", date_raw);
|
|
||||||
|
|
||||||
let date = NaiveDateTime::parse_from_str(&*date_add, "%d.%m.%Y %H:%M:%S");
|
|
||||||
|
|
||||||
Day {
|
|
||||||
name,
|
|
||||||
street: None,
|
|
||||||
date: date.unwrap().and_utc(),
|
|
||||||
lessons: Vec::new(),
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let lesson_time_column = days_markup[0].column + 1;
|
|
||||||
|
|
||||||
let row_distance = if day_index != days_markup.len() - 1 {
|
|
||||||
days_markup[day_index + 1].row
|
|
||||||
} else {
|
|
||||||
saturday_end_row
|
|
||||||
} - day_markup.row;
|
|
||||||
|
|
||||||
if days_times.len() != 6 {
|
|
||||||
let mut day_times: Vec<InternalTime> = Vec::new();
|
|
||||||
|
|
||||||
for row in day_markup.row..(day_markup.row + row_distance) {
|
|
||||||
// time
|
|
||||||
let time_opt = get_string_from_cell(&worksheet, row, lesson_time_column);
|
|
||||||
if time_opt.is_none() {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
let time = time_opt.unwrap();
|
|
||||||
|
|
||||||
// type
|
|
||||||
let lesson_type = if time.contains("пара") {
|
|
||||||
LessonType::Default
|
|
||||||
} else {
|
|
||||||
LessonType::Additional
|
|
||||||
};
|
|
||||||
|
|
||||||
// lesson index
|
|
||||||
let default_index = if lesson_type == LessonType::Default {
|
|
||||||
Some(
|
|
||||||
time.chars()
|
|
||||||
.next()
|
|
||||||
.unwrap()
|
|
||||||
.to_string()
|
|
||||||
.parse::<u32>()
|
|
||||||
.unwrap(),
|
|
||||||
)
|
|
||||||
} else {
|
|
||||||
None
|
|
||||||
};
|
|
||||||
|
|
||||||
// time
|
|
||||||
let time_range = {
|
|
||||||
static TIME_RE: LazyLock<Regex> =
|
|
||||||
LazyLock::new(|| Regex::new(r"(\d+\.\d+)-(\d+\.\d+)").unwrap());
|
|
||||||
|
|
||||||
let parse_res = TIME_RE.captures(&time).ok_or(ParseError::GlobalTime(
|
|
||||||
ErrorCell::new(row, lesson_time_column, time.clone()),
|
|
||||||
))?;
|
|
||||||
|
|
||||||
let start_match = parse_res.get(1).unwrap().as_str();
|
|
||||||
let start_parts: Vec<&str> = start_match.split(".").collect();
|
|
||||||
|
|
||||||
let end_match = parse_res.get(2).unwrap().as_str();
|
|
||||||
let end_parts: Vec<&str> = end_match.split(".").collect();
|
|
||||||
|
|
||||||
static GET_TIME: fn(DateTime<Utc>, &Vec<&str>) -> DateTime<Utc> =
|
|
||||||
|date, parts| {
|
|
||||||
date + Duration::hours(parts[0].parse::<i64>().unwrap() - 4)
|
|
||||||
+ Duration::minutes(parts[1].parse::<i64>().unwrap())
|
|
||||||
};
|
|
||||||
|
|
||||||
LessonTime {
|
|
||||||
start: GET_TIME(day.date.clone(), &start_parts),
|
|
||||||
end: GET_TIME(day.date.clone(), &end_parts),
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
day_times.push(InternalTime {
|
|
||||||
time_range,
|
|
||||||
lesson_type,
|
|
||||||
default_index,
|
|
||||||
xls_range: get_merge_from_start(&worksheet, row, lesson_time_column),
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
days_times.push(day_times);
|
|
||||||
}
|
|
||||||
|
|
||||||
let day_times = &days_times[day_index];
|
|
||||||
|
|
||||||
for time in day_times {
|
|
||||||
match &mut parse_lesson(
|
|
||||||
&worksheet,
|
|
||||||
&mut day,
|
|
||||||
&day_times,
|
|
||||||
&time,
|
|
||||||
group_markup.column,
|
|
||||||
)? {
|
|
||||||
Lessons(l) => day.lessons.append(l),
|
|
||||||
Street(s) => day.street = Some(s.to_owned()),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
group.days.push(day);
|
|
||||||
}
|
|
||||||
|
|
||||||
groups.insert(group.name.clone(), group);
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(ParseResult {
|
|
||||||
teachers: convert_groups_to_teachers(&groups),
|
|
||||||
groups,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
#[cfg(test)]
|
|
||||||
pub mod tests {
|
|
||||||
use super::*;
|
|
||||||
|
|
||||||
pub fn test_result() -> Result<ParseResult, ParseError> {
|
|
||||||
parse_xls(&include_bytes!("../../schedule.xls").to_vec())
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn read() {
|
|
||||||
let result = test_result();
|
|
||||||
|
|
||||||
assert!(result.is_ok());
|
|
||||||
|
|
||||||
assert_ne!(result.as_ref().unwrap().groups.len(), 0);
|
|
||||||
assert_ne!(result.as_ref().unwrap().teachers.len(), 0);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -150,7 +150,7 @@ mod tests {
|
|||||||
use std::fmt::Write;
|
use std::fmt::Write;
|
||||||
|
|
||||||
async fn sign_in_client(data: Request) -> ServiceResponse {
|
async fn sign_in_client(data: Request) -> ServiceResponse {
|
||||||
let app = test_app(test_app_state().await, sign_in).await;
|
let app = test_app(test_app_state(Default::default()).await, sign_in).await;
|
||||||
|
|
||||||
let req = test::TestRequest::with_uri("/sign-in")
|
let req = test::TestRequest::with_uri("/sign-in")
|
||||||
.method(Method::POST)
|
.method(Method::POST)
|
||||||
|
|||||||
@@ -241,7 +241,9 @@ mod tests {
|
|||||||
use crate::database::models::UserRole;
|
use crate::database::models::UserRole;
|
||||||
use crate::routes::auth::sign_up::schema::Request;
|
use crate::routes::auth::sign_up::schema::Request;
|
||||||
use crate::routes::auth::sign_up::sign_up;
|
use crate::routes::auth::sign_up::sign_up;
|
||||||
use crate::test_env::tests::{static_app_state, test_app_state, test_env};
|
use crate::test_env::tests::{
|
||||||
|
TestAppStateParams, TestScheduleType, static_app_state, test_app_state, test_env,
|
||||||
|
};
|
||||||
use actix_test::test_app;
|
use actix_test::test_app;
|
||||||
use actix_web::dev::ServiceResponse;
|
use actix_web::dev::ServiceResponse;
|
||||||
use actix_web::http::Method;
|
use actix_web::http::Method;
|
||||||
@@ -252,10 +254,22 @@ mod tests {
|
|||||||
username: String,
|
username: String,
|
||||||
group: String,
|
group: String,
|
||||||
role: UserRole,
|
role: UserRole,
|
||||||
|
load_schedule: bool,
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn sign_up_client(data: SignUpPartial) -> ServiceResponse {
|
async fn sign_up_client(data: SignUpPartial) -> ServiceResponse {
|
||||||
let app = test_app(test_app_state().await, sign_up).await;
|
let app = test_app(
|
||||||
|
test_app_state(TestAppStateParams {
|
||||||
|
schedule: if data.load_schedule {
|
||||||
|
TestScheduleType::Local
|
||||||
|
} else {
|
||||||
|
TestScheduleType::None
|
||||||
|
},
|
||||||
|
})
|
||||||
|
.await,
|
||||||
|
sign_up,
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
|
||||||
let req = test::TestRequest::with_uri("/sign-up")
|
let req = test::TestRequest::with_uri("/sign-up")
|
||||||
.method(Method::POST)
|
.method(Method::POST)
|
||||||
@@ -286,6 +300,7 @@ mod tests {
|
|||||||
username: "test::sign_up_valid".to_string(),
|
username: "test::sign_up_valid".to_string(),
|
||||||
group: "ИС-214/23".to_string(),
|
group: "ИС-214/23".to_string(),
|
||||||
role: UserRole::Student,
|
role: UserRole::Student,
|
||||||
|
load_schedule: false,
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -305,6 +320,7 @@ mod tests {
|
|||||||
username: "test::sign_up_multiple".to_string(),
|
username: "test::sign_up_multiple".to_string(),
|
||||||
group: "ИС-214/23".to_string(),
|
group: "ИС-214/23".to_string(),
|
||||||
role: UserRole::Student,
|
role: UserRole::Student,
|
||||||
|
load_schedule: false,
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -314,6 +330,7 @@ mod tests {
|
|||||||
username: "test::sign_up_multiple".to_string(),
|
username: "test::sign_up_multiple".to_string(),
|
||||||
group: "ИС-214/23".to_string(),
|
group: "ИС-214/23".to_string(),
|
||||||
role: UserRole::Student,
|
role: UserRole::Student,
|
||||||
|
load_schedule: false,
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -329,6 +346,7 @@ mod tests {
|
|||||||
username: "test::sign_up_invalid_role".to_string(),
|
username: "test::sign_up_invalid_role".to_string(),
|
||||||
group: "ИС-214/23".to_string(),
|
group: "ИС-214/23".to_string(),
|
||||||
role: UserRole::Admin,
|
role: UserRole::Admin,
|
||||||
|
load_schedule: false,
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
@@ -344,6 +362,7 @@ mod tests {
|
|||||||
username: "test::sign_up_invalid_group".to_string(),
|
username: "test::sign_up_invalid_group".to_string(),
|
||||||
group: "invalid_group".to_string(),
|
group: "invalid_group".to_string(),
|
||||||
role: UserRole::Student,
|
role: UserRole::Student,
|
||||||
|
load_schedule: true,
|
||||||
})
|
})
|
||||||
.await;
|
.await;
|
||||||
|
|
||||||
|
|||||||
@@ -39,7 +39,7 @@ pub async fn group(user: SyncExtractor<User>, app_state: web::Data<AppState>) ->
|
|||||||
}
|
}
|
||||||
|
|
||||||
mod schema {
|
mod schema {
|
||||||
use crate::parser::schema::ScheduleEntry;
|
use schedule_parser::schema::ScheduleEntry;
|
||||||
use actix_macros::{IntoResponseErrorNamed, StatusCode};
|
use actix_macros::{IntoResponseErrorNamed, StatusCode};
|
||||||
use chrono::{DateTime, NaiveDateTime, Utc};
|
use chrono::{DateTime, NaiveDateTime, Utc};
|
||||||
use derive_more::Display;
|
use derive_more::Display;
|
||||||
|
|||||||
@@ -2,9 +2,9 @@ mod cache_status;
|
|||||||
mod group;
|
mod group;
|
||||||
mod group_names;
|
mod group_names;
|
||||||
mod schedule;
|
mod schedule;
|
||||||
|
mod schema;
|
||||||
mod teacher;
|
mod teacher;
|
||||||
mod teacher_names;
|
mod teacher_names;
|
||||||
mod schema;
|
|
||||||
mod update_download_url;
|
mod update_download_url;
|
||||||
|
|
||||||
pub use cache_status::*;
|
pub use cache_status::*;
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
use crate::app_state::{AppState, Schedule};
|
use crate::app_state::{AppState, Schedule};
|
||||||
use crate::parser::schema::ScheduleEntry;
|
use schedule_parser::schema::ScheduleEntry;
|
||||||
use actix_macros::{IntoResponseErrorNamed, ResponderJson, StatusCode};
|
use actix_macros::{IntoResponseErrorNamed, ResponderJson, StatusCode};
|
||||||
use actix_web::web;
|
use actix_web::web;
|
||||||
use chrono::{DateTime, Duration, Utc};
|
use chrono::{DateTime, Duration, Utc};
|
||||||
@@ -99,7 +99,7 @@ impl From<&Schedule> for CacheStatus {
|
|||||||
fn from(value: &Schedule) -> Self {
|
fn from(value: &Schedule) -> Self {
|
||||||
Self {
|
Self {
|
||||||
cache_hash: value.hash(),
|
cache_hash: value.hash(),
|
||||||
cache_update_required: (value.fetched_at - Utc::now()) > Duration::minutes(5),
|
cache_update_required: (Utc::now() - value.fetched_at) > Duration::minutes(5),
|
||||||
last_cache_update: value.fetched_at.timestamp(),
|
last_cache_update: value.fetched_at.timestamp(),
|
||||||
last_schedule_update: value.updated_at.timestamp(),
|
last_schedule_update: value.updated_at.timestamp(),
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -40,7 +40,7 @@ pub async fn teacher(
|
|||||||
}
|
}
|
||||||
|
|
||||||
mod schema {
|
mod schema {
|
||||||
use crate::parser::schema::ScheduleEntry;
|
use schedule_parser::schema::ScheduleEntry;
|
||||||
use actix_macros::{IntoResponseErrorNamed, StatusCode};
|
use actix_macros::{IntoResponseErrorNamed, StatusCode};
|
||||||
use chrono::{DateTime, NaiveDateTime, Utc};
|
use chrono::{DateTime, NaiveDateTime, Utc};
|
||||||
use derive_more::Display;
|
use derive_more::Display;
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
use self::schema::*;
|
use self::schema::*;
|
||||||
use crate::AppState;
|
use crate::AppState;
|
||||||
use crate::app_state::Schedule;
|
use crate::app_state::Schedule;
|
||||||
use crate::parser::parse_xls;
|
use schedule_parser::parse_xls;
|
||||||
use crate::routes::schedule::schema::CacheStatus;
|
use crate::routes::schedule::schema::CacheStatus;
|
||||||
use crate::routes::schema::{IntoResponseAsError, ResponseError};
|
use crate::routes::schema::{IntoResponseAsError, ResponseError};
|
||||||
use crate::xls_downloader::interface::{FetchError, XLSDownloader};
|
use crate::xls_downloader::interface::{FetchError, XLSDownloader};
|
||||||
@@ -60,31 +60,32 @@ pub async fn update_download_url(
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
Err(error) => {
|
Err(error) => {
|
||||||
if let FetchError::Unknown(error) = error {
|
if let FetchError::Unknown(error) = &error {
|
||||||
sentry::capture_error(&error);
|
sentry::capture_error(&error);
|
||||||
}
|
}
|
||||||
|
|
||||||
ErrorCode::DownloadFailed.into_response()
|
ErrorCode::DownloadFailed(error).into_response()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Err(error) => {
|
Err(error) => {
|
||||||
if let FetchError::Unknown(error) = error {
|
if let FetchError::Unknown(error) = &error {
|
||||||
sentry::capture_error(&error);
|
sentry::capture_error(&error);
|
||||||
}
|
}
|
||||||
|
|
||||||
ErrorCode::FetchFailed.into_response()
|
ErrorCode::FetchFailed(error).into_response()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
mod schema {
|
mod schema {
|
||||||
use crate::parser::schema::ParseError;
|
use schedule_parser::schema::ParseError;
|
||||||
use crate::routes::schedule::schema::CacheStatus;
|
use crate::routes::schedule::schema::CacheStatus;
|
||||||
use actix_macros::{IntoResponseErrorNamed, StatusCode};
|
use actix_macros::{IntoResponseErrorNamed, StatusCode};
|
||||||
use derive_more::Display;
|
use derive_more::Display;
|
||||||
use serde::{Deserialize, Serialize, Serializer};
|
use serde::{Deserialize, Serialize, Serializer};
|
||||||
use utoipa::ToSchema;
|
use utoipa::ToSchema;
|
||||||
|
use crate::xls_downloader::interface::FetchError;
|
||||||
|
|
||||||
pub type ServiceResponse = crate::routes::schema::Response<CacheStatus, ErrorCode>;
|
pub type ServiceResponse = crate::routes::schema::Response<CacheStatus, ErrorCode>;
|
||||||
|
|
||||||
@@ -103,12 +104,12 @@ mod schema {
|
|||||||
NonWhitelistedHost,
|
NonWhitelistedHost,
|
||||||
|
|
||||||
/// Failed to retrieve file metadata.
|
/// Failed to retrieve file metadata.
|
||||||
#[display("Unable to retrieve metadata from the specified URL.")]
|
#[display("Unable to retrieve metadata from the specified URL: {_0}")]
|
||||||
FetchFailed,
|
FetchFailed(FetchError),
|
||||||
|
|
||||||
/// Failed to download the file.
|
/// Failed to download the file.
|
||||||
#[display("Unable to retrieve data from the specified URL.")]
|
#[display("Unable to retrieve data from the specified URL: {_0}")]
|
||||||
DownloadFailed,
|
DownloadFailed(FetchError),
|
||||||
|
|
||||||
/// The link leads to an outdated schedule.
|
/// The link leads to an outdated schedule.
|
||||||
///
|
///
|
||||||
@@ -129,8 +130,8 @@ mod schema {
|
|||||||
{
|
{
|
||||||
match self {
|
match self {
|
||||||
ErrorCode::NonWhitelistedHost => serializer.serialize_str("NON_WHITELISTED_HOST"),
|
ErrorCode::NonWhitelistedHost => serializer.serialize_str("NON_WHITELISTED_HOST"),
|
||||||
ErrorCode::FetchFailed => serializer.serialize_str("FETCH_FAILED"),
|
ErrorCode::FetchFailed(_) => serializer.serialize_str("FETCH_FAILED"),
|
||||||
ErrorCode::DownloadFailed => serializer.serialize_str("DOWNLOAD_FAILED"),
|
ErrorCode::DownloadFailed(_) => serializer.serialize_str("DOWNLOAD_FAILED"),
|
||||||
ErrorCode::OutdatedSchedule => serializer.serialize_str("OUTDATED_SCHEDULE"),
|
ErrorCode::OutdatedSchedule => serializer.serialize_str("OUTDATED_SCHEDULE"),
|
||||||
ErrorCode::InvalidSchedule(_) => serializer.serialize_str("INVALID_SCHEDULE"),
|
ErrorCode::InvalidSchedule(_) => serializer.serialize_str("INVALID_SCHEDULE"),
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,24 +1,47 @@
|
|||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
pub(crate) mod tests {
|
pub(crate) mod tests {
|
||||||
use crate::app_state::{AppState, Schedule, app_state};
|
use crate::app_state::{AppState, Schedule, app_state};
|
||||||
use crate::parser::tests::test_result;
|
use schedule_parser::test_utils::test_result;
|
||||||
|
use crate::utility::mutex::MutexScope;
|
||||||
use actix_web::web;
|
use actix_web::web;
|
||||||
|
use std::default::Default;
|
||||||
use tokio::sync::OnceCell;
|
use tokio::sync::OnceCell;
|
||||||
|
|
||||||
pub fn test_env() {
|
pub fn test_env() {
|
||||||
dotenvy::from_path(".env.test").expect("Failed to load test environment file");
|
dotenvy::from_path(".env.test").expect("Failed to load test environment file");
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn test_app_state() -> web::Data<AppState> {
|
pub enum TestScheduleType {
|
||||||
let state = app_state().await;
|
None,
|
||||||
let mut schedule_lock = state.schedule.lock().unwrap();
|
Local,
|
||||||
|
}
|
||||||
|
|
||||||
*schedule_lock = Some(Schedule {
|
pub struct TestAppStateParams {
|
||||||
etag: "".to_string(),
|
pub schedule: TestScheduleType,
|
||||||
fetched_at: Default::default(),
|
}
|
||||||
updated_at: Default::default(),
|
|
||||||
parsed_at: Default::default(),
|
impl Default for TestAppStateParams {
|
||||||
data: test_result().unwrap(),
|
fn default() -> Self {
|
||||||
|
Self {
|
||||||
|
schedule: TestScheduleType::None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn test_app_state(params: TestAppStateParams) -> web::Data<AppState> {
|
||||||
|
let state = app_state().await;
|
||||||
|
|
||||||
|
state.schedule.scope(|schedule| {
|
||||||
|
*schedule = match params.schedule {
|
||||||
|
TestScheduleType::None => None,
|
||||||
|
TestScheduleType::Local => Some(Schedule {
|
||||||
|
etag: "".to_string(),
|
||||||
|
fetched_at: Default::default(),
|
||||||
|
updated_at: Default::default(),
|
||||||
|
parsed_at: Default::default(),
|
||||||
|
data: test_result().unwrap(),
|
||||||
|
}),
|
||||||
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
state.clone()
|
state.clone()
|
||||||
@@ -27,6 +50,9 @@ pub(crate) mod tests {
|
|||||||
pub async fn static_app_state() -> web::Data<AppState> {
|
pub async fn static_app_state() -> web::Data<AppState> {
|
||||||
static STATE: OnceCell<web::Data<AppState>> = OnceCell::const_new();
|
static STATE: OnceCell<web::Data<AppState>> = OnceCell::const_new();
|
||||||
|
|
||||||
STATE.get_or_init(|| test_app_state()).await.clone()
|
STATE
|
||||||
|
.get_or_init(|| test_app_state(Default::default()))
|
||||||
|
.await
|
||||||
|
.clone()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
use crate::xls_downloader::interface::{FetchError, FetchOk, FetchResult, XLSDownloader};
|
use crate::xls_downloader::interface::{FetchError, FetchOk, FetchResult, XLSDownloader};
|
||||||
use chrono::{DateTime, Utc};
|
use chrono::{DateTime, Utc};
|
||||||
use std::env;
|
use std::env;
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
pub struct BasicXlsDownloader {
|
pub struct BasicXlsDownloader {
|
||||||
pub url: Option<String>,
|
pub url: Option<String>,
|
||||||
@@ -22,7 +23,7 @@ async fn fetch_specified(url: &String, user_agent: &String, head: bool) -> Fetch
|
|||||||
match response {
|
match response {
|
||||||
Ok(r) => {
|
Ok(r) => {
|
||||||
if r.status().as_u16() != 200 {
|
if r.status().as_u16() != 200 {
|
||||||
return Err(FetchError::BadStatusCode);
|
return Err(FetchError::BadStatusCode(r.status().as_u16()));
|
||||||
}
|
}
|
||||||
|
|
||||||
let headers = r.headers();
|
let headers = r.headers();
|
||||||
@@ -32,11 +33,18 @@ async fn fetch_specified(url: &String, user_agent: &String, head: bool) -> Fetch
|
|||||||
let last_modified = headers.get("last-modified");
|
let last_modified = headers.get("last-modified");
|
||||||
let date = headers.get("date");
|
let date = headers.get("date");
|
||||||
|
|
||||||
if content_type.is_none() || etag.is_none() || last_modified.is_none() || date.is_none()
|
if content_type.is_none() {
|
||||||
{
|
Err(FetchError::BadHeaders("Content-Type".to_string()))
|
||||||
Err(FetchError::BadHeaders)
|
} else if etag.is_none() {
|
||||||
|
Err(FetchError::BadHeaders("ETag".to_string()))
|
||||||
|
} else if last_modified.is_none() {
|
||||||
|
Err(FetchError::BadHeaders("Last-Modified".to_string()))
|
||||||
|
} else if date.is_none() {
|
||||||
|
Err(FetchError::BadHeaders("Date".to_string()))
|
||||||
} else if content_type.unwrap() != "application/vnd.ms-excel" {
|
} else if content_type.unwrap() != "application/vnd.ms-excel" {
|
||||||
Err(FetchError::BadContentType)
|
Err(FetchError::BadContentType(
|
||||||
|
content_type.unwrap().to_str().unwrap().to_string(),
|
||||||
|
))
|
||||||
} else {
|
} else {
|
||||||
let etag = etag.unwrap().to_str().unwrap().to_string();
|
let etag = etag.unwrap().to_str().unwrap().to_string();
|
||||||
let last_modified =
|
let last_modified =
|
||||||
@@ -51,7 +59,7 @@ async fn fetch_specified(url: &String, user_agent: &String, head: bool) -> Fetch
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Err(e) => Err(FetchError::Unknown(e)),
|
Err(error) => Err(FetchError::Unknown(Arc::new(error))),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -86,7 +94,7 @@ impl XLSDownloader for BasicXlsDownloader {
|
|||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use crate::xls_downloader::basic_impl::{BasicXlsDownloader, fetch_specified};
|
use crate::xls_downloader::basic_impl::{fetch_specified, BasicXlsDownloader};
|
||||||
use crate::xls_downloader::interface::{FetchError, XLSDownloader};
|
use crate::xls_downloader::interface::{FetchError, XLSDownloader};
|
||||||
|
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
@@ -116,14 +124,10 @@ mod tests {
|
|||||||
assert!(results[0].is_err());
|
assert!(results[0].is_err());
|
||||||
assert!(results[1].is_err());
|
assert!(results[1].is_err());
|
||||||
|
|
||||||
assert_eq!(
|
let expected_error = FetchError::BadStatusCode(404);
|
||||||
*results[0].as_ref().err().unwrap(),
|
|
||||||
FetchError::BadStatusCode
|
assert_eq!(*results[0].as_ref().err().unwrap(), expected_error);
|
||||||
);
|
assert_eq!(*results[1].as_ref().err().unwrap(), expected_error);
|
||||||
assert_eq!(
|
|
||||||
*results[1].as_ref().err().unwrap(),
|
|
||||||
FetchError::BadStatusCode
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
@@ -139,8 +143,10 @@ mod tests {
|
|||||||
assert!(results[0].is_err());
|
assert!(results[0].is_err());
|
||||||
assert!(results[1].is_err());
|
assert!(results[1].is_err());
|
||||||
|
|
||||||
assert_eq!(*results[0].as_ref().err().unwrap(), FetchError::BadHeaders);
|
let expected_error = FetchError::BadHeaders("ETag".to_string());
|
||||||
assert_eq!(*results[1].as_ref().err().unwrap(), FetchError::BadHeaders);
|
|
||||||
|
assert_eq!(*results[0].as_ref().err().unwrap(), expected_error);
|
||||||
|
assert_eq!(*results[1].as_ref().err().unwrap(), expected_error);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
@@ -155,15 +161,6 @@ mod tests {
|
|||||||
|
|
||||||
assert!(results[0].is_err());
|
assert!(results[0].is_err());
|
||||||
assert!(results[1].is_err());
|
assert!(results[1].is_err());
|
||||||
|
|
||||||
assert_eq!(
|
|
||||||
*results[0].as_ref().err().unwrap(),
|
|
||||||
FetchError::BadContentType
|
|
||||||
);
|
|
||||||
assert_eq!(
|
|
||||||
*results[1].as_ref().err().unwrap(),
|
|
||||||
FetchError::BadContentType
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
|
|||||||
@@ -1,23 +1,32 @@
|
|||||||
use chrono::{DateTime, Utc};
|
use chrono::{DateTime, Utc};
|
||||||
|
use derive_more::Display;
|
||||||
use std::mem::discriminant;
|
use std::mem::discriminant;
|
||||||
|
use std::sync::Arc;
|
||||||
|
use utoipa::ToSchema;
|
||||||
|
|
||||||
/// XLS data retrieval errors.
|
/// XLS data retrieval errors.
|
||||||
#[derive(Debug)]
|
#[derive(Clone, Debug, ToSchema, Display)]
|
||||||
pub enum FetchError {
|
pub enum FetchError {
|
||||||
/// File url is not set.
|
/// File url is not set.
|
||||||
|
#[display("The link to the timetable was not provided earlier.")]
|
||||||
NoUrlProvided,
|
NoUrlProvided,
|
||||||
|
|
||||||
/// Unknown error.
|
/// Unknown error.
|
||||||
Unknown(reqwest::Error),
|
#[display("An unknown error occurred while downloading the file.")]
|
||||||
|
#[schema(value_type = String)]
|
||||||
|
Unknown(Arc<reqwest::Error>),
|
||||||
|
|
||||||
/// Server returned a status code different from 200.
|
/// Server returned a status code different from 200.
|
||||||
BadStatusCode,
|
#[display("Server returned a status code {_0}.")]
|
||||||
|
BadStatusCode(u16),
|
||||||
|
|
||||||
/// The url leads to a file of a different type.
|
/// The url leads to a file of a different type.
|
||||||
BadContentType,
|
#[display("The link leads to a file of type '{_0}'.")]
|
||||||
|
BadContentType(String),
|
||||||
|
|
||||||
/// Server doesn't return expected headers.
|
/// Server doesn't return expected headers.
|
||||||
BadHeaders,
|
#[display("Server doesn't return expected header(s) '{_0}'.")]
|
||||||
|
BadHeaders(String),
|
||||||
}
|
}
|
||||||
|
|
||||||
impl PartialEq for FetchError {
|
impl PartialEq for FetchError {
|
||||||
|
|||||||
Reference in New Issue
Block a user