mirror of
https://github.com/n08i40k/schedule-parser-rusted.git
synced 2025-12-06 09:47:50 +03:00
Compare commits
18 Commits
57c1699c9a
...
release/v1
| Author | SHA1 | Date | |
|---|---|---|---|
|
b664ba578d
|
|||
|
983967f8b0
|
|||
|
e5760120e2
|
|||
|
a28fb66dd4
|
|||
|
3780fb3136
|
|||
|
6c71bc19f5
|
|||
|
2d0041dc8b
|
|||
|
b5d372e109
|
|||
|
84dca02c34
|
|||
|
6c9d3b3b31
|
|||
|
a348b1b99b
|
|||
|
ff12ee5da2
|
|||
|
35f707901f
|
|||
|
edea6c5424
|
|||
|
fdbb872fc3
|
|||
|
dbc800fef1
|
|||
|
e729d84c93
|
|||
|
cc7adf10ed
|
142
.github/workflows/build.yml
vendored
Normal file
142
.github/workflows/build.yml
vendored
Normal file
@@ -0,0 +1,142 @@
|
||||
name: build
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [ "master" ]
|
||||
tags-ignore: [ "release/v*" ]
|
||||
|
||||
permissions:
|
||||
contents: write
|
||||
|
||||
env:
|
||||
CARGO_TERM_COLOR: always
|
||||
|
||||
BINARY_NAME: schedule-parser-rusted
|
||||
|
||||
TEST_DB: ${{ secrets.TEST_DATABASE_URL }}
|
||||
|
||||
SENTRY_AUTH_TOKEN: ${{ secrets.SENTRY_AUTH_TOKEN }}
|
||||
SENTRY_ORG: ${{ secrets.SENTRY_ORG }}
|
||||
SENTRY_PROJECT: ${{ secrets.SENTRY_PROJECT }}
|
||||
|
||||
DOCKER_IMAGE_NAME: ${{ github.repository }}
|
||||
|
||||
DOCKER_REGISTRY_HOST: registry.n08i40k.ru
|
||||
DOCKER_REGISTRY_USERNAME: ${{ github.repository_owner }}
|
||||
DOCKER_REGISTRY_PASSWORD: ${{ secrets.DOCKER_REGISTRY_PASSWORD }}
|
||||
|
||||
jobs:
|
||||
test:
|
||||
name: Test
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Setup Rust
|
||||
uses: actions-rust-lang/setup-rust-toolchain@v1.11.0
|
||||
with:
|
||||
toolchain: stable
|
||||
|
||||
- name: Test
|
||||
run: |
|
||||
cargo test --verbose
|
||||
env:
|
||||
DATABASE_URL: ${{ env.TEST_DB }}
|
||||
SCHEDULE_DISABLE_AUTO_UPDATE: 1
|
||||
JWT_SECRET: "test-secret-at-least-256-bits-used"
|
||||
VK_ID_CLIENT_ID: 0
|
||||
VK_ID_REDIRECT_URI: "vk0://vk.com/blank.html"
|
||||
TELEGRAM_BOT_ID: 0
|
||||
TELEGRAM_MINI_APP_HOST: example.com
|
||||
TELEGRAM_TEST_DC: false
|
||||
YANDEX_CLOUD_API_KEY: ""
|
||||
YANDEX_CLOUD_FUNC_ID: ""
|
||||
build:
|
||||
name: Build
|
||||
runs-on: ubuntu-latest
|
||||
needs: test
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Setup Rust
|
||||
uses: actions-rust-lang/setup-rust-toolchain@v1.11.0
|
||||
with:
|
||||
toolchain: stable
|
||||
|
||||
- name: Build
|
||||
run: cargo build --release --verbose
|
||||
|
||||
- name: Extract debug symbols
|
||||
run: |
|
||||
objcopy --only-keep-debug target/release/${{ env.BINARY_NAME }}{,.d}
|
||||
objcopy --strip-debug --strip-unneeded target/release/${{ env.BINARY_NAME }}
|
||||
objcopy --add-gnu-debuglink target/release/${{ env.BINARY_NAME }}{.d,}
|
||||
|
||||
- name: Setup sentry-cli
|
||||
uses: matbour/setup-sentry-cli@v2.0.0
|
||||
with:
|
||||
version: latest
|
||||
token: ${{ env.SENTRY_AUTH_TOKEN }}
|
||||
organization: ${{ env.SENTRY_ORG }}
|
||||
project: ${{ env.SENTRY_PROJECT }}
|
||||
|
||||
- name: Upload debug symbols to Sentry
|
||||
run: |
|
||||
sentry-cli debug-files upload --include-sources .
|
||||
|
||||
- name: Upload build binary artifact
|
||||
uses: actions/upload-artifact@v4
|
||||
with:
|
||||
name: release-binary
|
||||
path: target/release/${{ env.BINARY_NAME }}
|
||||
|
||||
- name: Upload build debug symbols artifact
|
||||
uses: actions/upload-artifact@v4
|
||||
with:
|
||||
name: release-symbols
|
||||
path: target/release/${{ env.BINARY_NAME }}.d
|
||||
|
||||
docker:
|
||||
name: Build & Push Docker Image
|
||||
runs-on: ubuntu-latest
|
||||
needs: build
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
|
||||
- name: Download build artifacts
|
||||
uses: actions/download-artifact@v4
|
||||
with:
|
||||
name: release-binary
|
||||
|
||||
- name: Setup Docker Buildx
|
||||
uses: docker/setup-buildx-action@v3.10.0
|
||||
|
||||
- name: Login to Registry
|
||||
uses: docker/login-action@v3.4.0
|
||||
with:
|
||||
registry: ${{ env.DOCKER_REGISTRY_HOST }}
|
||||
username: ${{ env.DOCKER_REGISTRY_USERNAME }}
|
||||
password: ${{ env.DOCKER_REGISTRY_PASSWORD }}
|
||||
|
||||
- name: Extract Docker metadata
|
||||
id: meta
|
||||
uses: docker/metadata-action@v5.7.0
|
||||
with:
|
||||
images: ${{ env.DOCKER_REGISTRY_HOST }}/${{ env.DOCKER_IMAGE_NAME }}
|
||||
|
||||
- name: Build and push Docker image
|
||||
id: build-and-push
|
||||
uses: docker/build-push-action@v6.15.0
|
||||
with:
|
||||
context: .
|
||||
push: true
|
||||
tags: ${{ steps.meta.outputs.tags }}
|
||||
labels: ${{ steps.meta.outputs.labels }}
|
||||
cache-from: type=gha
|
||||
cache-to: type=gha,mode=max
|
||||
build-args: |
|
||||
"BINARY_NAME=${{ env.BINARY_NAME }}"
|
||||
2
.github/workflows/test.yml
vendored
2
.github/workflows/test.yml
vendored
@@ -2,7 +2,7 @@ name: cargo test
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [ "master" ]
|
||||
branches: [ "development" ]
|
||||
tags-ignore: [ "release/v*" ]
|
||||
|
||||
permissions:
|
||||
|
||||
2872
Cargo.lock
generated
2872
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
30
Cargo.toml
30
Cargo.toml
@@ -3,13 +3,16 @@ members = ["actix-macros", "actix-test", "providers"]
|
||||
|
||||
[package]
|
||||
name = "schedule-parser-rusted"
|
||||
version = "1.1.1"
|
||||
version = "1.3.0"
|
||||
edition = "2024"
|
||||
publish = false
|
||||
|
||||
[profile.release]
|
||||
debug = true
|
||||
|
||||
[features]
|
||||
trace = ["tracing", "console-subscriber"]
|
||||
|
||||
[dependencies]
|
||||
providers = { path = "providers" }
|
||||
actix-macros = { path = "actix-macros" }
|
||||
@@ -18,13 +21,12 @@ actix-macros = { path = "actix-macros" }
|
||||
actix-web = "4.11.0"
|
||||
|
||||
# basic
|
||||
chrono = { version = "0.4.41", features = ["serde"] }
|
||||
chrono = { version = "0.4.42", features = ["serde"] }
|
||||
derive_more = { version = "2.0.1", features = ["full"] }
|
||||
dotenvy = "0.15.7"
|
||||
|
||||
# sql
|
||||
diesel = { version = "2.2.12", features = ["postgres"] }
|
||||
diesel-derive-enum = { git = "https://github.com/Havunen/diesel-derive-enum.git", features = ["postgres"] }
|
||||
database = { path = "database" }
|
||||
|
||||
# logging
|
||||
env_logger = "0.11.8"
|
||||
@@ -46,13 +48,13 @@ reqwest = { version = "0.12.23", features = ["json"] }
|
||||
mime = "0.3.17"
|
||||
|
||||
# error handling
|
||||
sentry = "0.42.0"
|
||||
sentry-actix = "0.42.0"
|
||||
sentry = "0.43.0"
|
||||
sentry-actix = "0.43.0"
|
||||
|
||||
# [de]serializing
|
||||
serde = { version = "1.0.219", features = ["derive"] }
|
||||
serde_json = "1.0.143"
|
||||
serde_with = "3.14.0"
|
||||
serde = { version = "1", features = ["derive"] }
|
||||
serde_json = "1"
|
||||
serde_with = "3.14"
|
||||
|
||||
sha1 = "0.11.0-rc.2"
|
||||
|
||||
@@ -63,13 +65,17 @@ utoipa-actix-web = "0.1.2"
|
||||
|
||||
uuid = { version = "1.18.1", features = ["v4"] }
|
||||
hex-literal = "1"
|
||||
log = "0.4.27"
|
||||
log = "0.4.28"
|
||||
|
||||
# telegram webdata deciding and verify
|
||||
base64 = "0.22.1"
|
||||
percent-encoding = "2.3.2"
|
||||
ed25519-dalek = "3.0.0-pre.0"
|
||||
ed25519-dalek = "3.0.0-pre.1"
|
||||
|
||||
# development tracing
|
||||
console-subscriber = { version = "0.4.1", optional = true }
|
||||
tracing = { version = "0.1.41", optional = true }
|
||||
|
||||
[dev-dependencies]
|
||||
providers = { path = "providers", features = ["test"] }
|
||||
actix-test = { path = "actix-test" }
|
||||
actix-test = { path = "actix-test" }
|
||||
@@ -6,7 +6,7 @@ mod shared {
|
||||
use quote::{ToTokens, quote};
|
||||
use syn::{Attribute, DeriveInput};
|
||||
|
||||
pub fn find_status_code(attrs: &Vec<Attribute>) -> Option<proc_macro2::TokenStream> {
|
||||
pub fn find_status_code(attrs: &[Attribute]) -> Option<proc_macro2::TokenStream> {
|
||||
attrs
|
||||
.iter()
|
||||
.find_map(|attr| -> Option<proc_macro2::TokenStream> {
|
||||
@@ -41,14 +41,12 @@ mod shared {
|
||||
|
||||
let mut status_code_arms: Vec<proc_macro2::TokenStream> = variants
|
||||
.iter()
|
||||
.map(|v| -> Option<proc_macro2::TokenStream> {
|
||||
.filter_map(|v| -> Option<proc_macro2::TokenStream> {
|
||||
let status_code = find_status_code(&v.attrs)?;
|
||||
let variant_name = &v.ident;
|
||||
|
||||
Some(quote! { #name::#variant_name => #status_code, })
|
||||
})
|
||||
.filter(|v| v.is_some())
|
||||
.map(|v| v.unwrap())
|
||||
.collect();
|
||||
|
||||
if status_code_arms.len() < variants.len() {
|
||||
|
||||
11
database/Cargo.toml
Normal file
11
database/Cargo.toml
Normal file
@@ -0,0 +1,11 @@
|
||||
[package]
|
||||
name = "database"
|
||||
version = "0.1.0"
|
||||
edition = "2024"
|
||||
|
||||
[dependencies]
|
||||
migration = { path = "migration" }
|
||||
entity = { path = "entity" }
|
||||
sea-orm = { version = "2.0.0-rc.6", features = ["sqlx-postgres", "runtime-tokio"] }
|
||||
|
||||
paste = "1.0.15"
|
||||
1
database/entity/.gitignore
vendored
Normal file
1
database/entity/.gitignore
vendored
Normal file
@@ -0,0 +1 @@
|
||||
/target
|
||||
9
database/entity/Cargo.toml
Normal file
9
database/entity/Cargo.toml
Normal file
@@ -0,0 +1,9 @@
|
||||
[package]
|
||||
name = "entity"
|
||||
version = "0.1.0"
|
||||
edition = "2024"
|
||||
|
||||
[dependencies]
|
||||
sea-orm = "2.0.0-rc.6"
|
||||
serde = { version = "1.0.219", features = ["derive"] }
|
||||
utoipa = "5.4.0"
|
||||
6
database/entity/src/lib.rs
Normal file
6
database/entity/src/lib.rs
Normal file
@@ -0,0 +1,6 @@
|
||||
//! `SeaORM` Entity, @generated by sea-orm-codegen 1.1.12
|
||||
|
||||
pub mod prelude;
|
||||
|
||||
pub mod sea_orm_active_enums;
|
||||
pub mod user;
|
||||
3
database/entity/src/prelude.rs
Normal file
3
database/entity/src/prelude.rs
Normal file
@@ -0,0 +1,3 @@
|
||||
//! `SeaORM` Entity, @generated by sea-orm-codegen 1.1.12
|
||||
|
||||
pub use super::user::Entity as User;
|
||||
25
database/entity/src/sea_orm_active_enums.rs
Normal file
25
database/entity/src/sea_orm_active_enums.rs
Normal file
@@ -0,0 +1,25 @@
|
||||
//! `SeaORM` Entity, @generated by sea-orm-codegen 1.1.12
|
||||
|
||||
use sea_orm::entity::prelude::*;
|
||||
|
||||
#[derive(
|
||||
Debug,
|
||||
Clone,
|
||||
PartialEq,
|
||||
Eq,
|
||||
EnumIter,
|
||||
DeriveActiveEnum,
|
||||
:: serde :: Serialize,
|
||||
:: serde :: Deserialize,
|
||||
:: utoipa :: ToSchema,
|
||||
)]
|
||||
#[sea_orm(rs_type = "String", db_type = "Enum", enum_name = "user_role")]
|
||||
#[serde(rename_all = "SCREAMING_SNAKE_CASE")]
|
||||
pub enum UserRole {
|
||||
#[sea_orm(string_value = "student")]
|
||||
Student,
|
||||
#[sea_orm(string_value = "teacher")]
|
||||
Teacher,
|
||||
#[sea_orm(string_value = "admin")]
|
||||
Admin,
|
||||
}
|
||||
25
database/entity/src/user.rs
Normal file
25
database/entity/src/user.rs
Normal file
@@ -0,0 +1,25 @@
|
||||
//! `SeaORM` Entity, @generated by sea-orm-codegen 1.1.12
|
||||
|
||||
use super::sea_orm_active_enums::UserRole;
|
||||
use sea_orm::entity::prelude::*;
|
||||
|
||||
#[derive(Clone, Debug, PartialEq, DeriveEntityModel, Eq)]
|
||||
#[sea_orm(table_name = "user")]
|
||||
pub struct Model {
|
||||
#[sea_orm(primary_key, auto_increment = false)]
|
||||
pub id: String,
|
||||
#[sea_orm(unique)]
|
||||
pub username: String,
|
||||
pub password: Option<String>,
|
||||
pub vk_id: Option<i32>,
|
||||
pub group: Option<String>,
|
||||
pub role: UserRole,
|
||||
pub android_version: Option<String>,
|
||||
#[sea_orm(unique)]
|
||||
pub telegram_id: Option<i64>,
|
||||
}
|
||||
|
||||
#[derive(Copy, Clone, Debug, EnumIter, DeriveRelation)]
|
||||
pub enum Relation {}
|
||||
|
||||
impl ActiveModelBehavior for ActiveModel {}
|
||||
1
database/migration/.gitignore
vendored
Normal file
1
database/migration/.gitignore
vendored
Normal file
@@ -0,0 +1 @@
|
||||
/target
|
||||
22
database/migration/Cargo.toml
Normal file
22
database/migration/Cargo.toml
Normal file
@@ -0,0 +1,22 @@
|
||||
[package]
|
||||
name = "migration"
|
||||
version = "0.1.0"
|
||||
edition = "2021"
|
||||
publish = false
|
||||
|
||||
[lib]
|
||||
name = "migration"
|
||||
path = "src/lib.rs"
|
||||
|
||||
[dependencies]
|
||||
async-std = { version = "1", features = ["attributes", "tokio1"] }
|
||||
|
||||
[dependencies.sea-orm-migration]
|
||||
version = "2.0.0-rc.6"
|
||||
features = [
|
||||
# Enable at least one `ASYNC_RUNTIME` and `DATABASE_DRIVER` feature if you want to run migration via CLI.
|
||||
# View the list of supported features at https://www.sea-ql.org/SeaORM/docs/install-and-config/database-and-async-runtime.
|
||||
# e.g.
|
||||
"runtime-tokio", # `ASYNC_RUNTIME` feature
|
||||
"sqlx-postgres", # `DATABASE_DRIVER` feature
|
||||
]
|
||||
16
database/migration/src/lib.rs
Normal file
16
database/migration/src/lib.rs
Normal file
@@ -0,0 +1,16 @@
|
||||
pub use sea_orm_migration::prelude::MigratorTrait;
|
||||
|
||||
use sea_orm_migration::prelude::*;
|
||||
|
||||
mod m20250904_024854_init;
|
||||
|
||||
pub struct Migrator;
|
||||
|
||||
#[async_trait::async_trait]
|
||||
impl MigratorTrait for Migrator {
|
||||
fn migrations() -> Vec<Box<dyn MigrationTrait>> {
|
||||
vec![
|
||||
Box::new(m20250904_024854_init::Migration),
|
||||
]
|
||||
}
|
||||
}
|
||||
70
database/migration/src/m20250904_024854_init.rs
Normal file
70
database/migration/src/m20250904_024854_init.rs
Normal file
@@ -0,0 +1,70 @@
|
||||
use sea_orm_migration::prelude::extension::postgres::Type;
|
||||
use sea_orm_migration::sea_orm::{EnumIter, Iterable};
|
||||
use sea_orm_migration::{prelude::*, schema::*};
|
||||
|
||||
#[derive(DeriveMigrationName)]
|
||||
pub struct Migration;
|
||||
|
||||
#[async_trait::async_trait]
|
||||
impl MigrationTrait for Migration {
|
||||
async fn up(&self, manager: &SchemaManager) -> Result<(), DbErr> {
|
||||
manager
|
||||
.create_type(
|
||||
Type::create()
|
||||
.as_enum(UserRole)
|
||||
.values(UserRoleVariants::iter())
|
||||
.to_owned(),
|
||||
)
|
||||
.await?;
|
||||
|
||||
manager
|
||||
.create_table(
|
||||
Table::create()
|
||||
.table(User::Table)
|
||||
.if_not_exists()
|
||||
.col(string_uniq(User::Id).primary_key().not_null())
|
||||
.col(string_uniq(User::Username).not_null())
|
||||
.col(string_null(User::Password))
|
||||
.col(integer_null(User::VkId))
|
||||
.col(string_null(User::Group))
|
||||
.col(enumeration(User::Role, UserRole, UserRoleVariants::iter()))
|
||||
.col(string_null(User::AndroidVersion))
|
||||
.col(big_integer_null(User::TelegramId).unique_key())
|
||||
.to_owned(),
|
||||
)
|
||||
.await
|
||||
}
|
||||
|
||||
async fn down(&self, manager: &SchemaManager) -> Result<(), DbErr> {
|
||||
manager
|
||||
.drop_table(Table::drop().table(User::Table).to_owned())
|
||||
.await?;
|
||||
|
||||
manager
|
||||
.drop_type(Type::drop().name(UserRole).to_owned())
|
||||
.await
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(DeriveIden)]
|
||||
struct UserRole;
|
||||
|
||||
#[derive(DeriveIden, EnumIter)]
|
||||
enum UserRoleVariants {
|
||||
Student,
|
||||
Teacher,
|
||||
Admin,
|
||||
}
|
||||
|
||||
#[derive(DeriveIden)]
|
||||
enum User {
|
||||
Table,
|
||||
Id,
|
||||
Username,
|
||||
Password,
|
||||
VkId,
|
||||
Group,
|
||||
Role,
|
||||
AndroidVersion,
|
||||
TelegramId,
|
||||
}
|
||||
6
database/migration/src/main.rs
Normal file
6
database/migration/src/main.rs
Normal file
@@ -0,0 +1,6 @@
|
||||
use sea_orm_migration::prelude::*;
|
||||
|
||||
#[async_std::main]
|
||||
async fn main() {
|
||||
cli::run_cli(migration::Migrator).await;
|
||||
}
|
||||
10
database/src/lib.rs
Normal file
10
database/src/lib.rs
Normal file
@@ -0,0 +1,10 @@
|
||||
pub mod query;
|
||||
|
||||
pub use migration;
|
||||
pub use sea_orm;
|
||||
|
||||
pub mod entity {
|
||||
pub use entity::*;
|
||||
|
||||
pub use entity::user::{ActiveModel as ActiveUser, Model as User, Entity as UserEntity, Column as UserColumn};
|
||||
}
|
||||
63
database/src/query.rs
Normal file
63
database/src/query.rs
Normal file
@@ -0,0 +1,63 @@
|
||||
use paste::paste;
|
||||
use sea_orm::ColumnTrait;
|
||||
use sea_orm::EntityTrait;
|
||||
use sea_orm::QueryFilter;
|
||||
|
||||
pub struct Query;
|
||||
|
||||
macro_rules! ref_type {
|
||||
(String) => {
|
||||
&String
|
||||
};
|
||||
(str) => {
|
||||
&str
|
||||
};
|
||||
($other:ty) => {
|
||||
$other
|
||||
};
|
||||
}
|
||||
|
||||
macro_rules! define_is_exists {
|
||||
($entity: ident, $by: ident, $by_type: ident, $by_column: ident) => {
|
||||
paste! {
|
||||
pub async fn [<is_ $entity _exists_by_ $by>](
|
||||
db: &::sea_orm::DbConn,
|
||||
$by: ref_type!($by_type)
|
||||
) -> Result<bool, ::sea_orm::DbErr> {
|
||||
::entity::$entity::Entity::find()
|
||||
.filter(::entity::$entity::Column::$by_column.eq($by))
|
||||
.one(db)
|
||||
.await
|
||||
.map(|x| x.is_some())
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
macro_rules! define_find_by {
|
||||
($entity: ident, $by: ident, $by_type: ident, $by_column: ident) => {
|
||||
paste! {
|
||||
pub async fn [<find_ $entity _by_ $by>](
|
||||
db: &::sea_orm::DbConn,
|
||||
$by: ref_type!($by_type)
|
||||
) -> Result<Option<::entity::$entity::Model>, ::sea_orm::DbErr> {
|
||||
::entity::$entity::Entity::find()
|
||||
.filter(::entity::$entity::Column::$by_column.eq($by))
|
||||
.one(db)
|
||||
.await
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
impl Query {
|
||||
define_find_by!(user, id, str, Id);
|
||||
define_find_by!(user, telegram_id, i64, TelegramId);
|
||||
define_find_by!(user, vk_id, i32, VkId);
|
||||
define_find_by!(user, username, str, Username);
|
||||
|
||||
define_is_exists!(user, id, str, Id);
|
||||
define_is_exists!(user, username, str, Username);
|
||||
define_is_exists!(user, telegram_id, i64, TelegramId);
|
||||
define_is_exists!(user, vk_id, i32, VkId);
|
||||
}
|
||||
@@ -1,9 +0,0 @@
|
||||
# For documentation on how to configure this file,
|
||||
# see https://diesel.rs/guides/configuring-diesel-cli
|
||||
|
||||
[print_schema]
|
||||
file = "src/database/schema.rs"
|
||||
custom_type_derives = ["diesel::query_builder::QueryId", "Clone"]
|
||||
|
||||
[migrations_directory]
|
||||
dir = "./migrations"
|
||||
@@ -1,6 +0,0 @@
|
||||
-- This file was automatically created by Diesel to set up helper functions
|
||||
-- and other internal bookkeeping. This file is safe to edit, any future
|
||||
-- changes will be added to existing projects as new migrations.
|
||||
|
||||
DROP FUNCTION IF EXISTS diesel_manage_updated_at(_tbl regclass);
|
||||
DROP FUNCTION IF EXISTS diesel_set_updated_at();
|
||||
@@ -1,36 +0,0 @@
|
||||
-- This file was automatically created by Diesel to set up helper functions
|
||||
-- and other internal bookkeeping. This file is safe to edit, any future
|
||||
-- changes will be added to existing projects as new migrations.
|
||||
|
||||
|
||||
|
||||
|
||||
-- Sets up a trigger for the given table to automatically set a column called
|
||||
-- `updated_at` whenever the row is modified (unless `updated_at` was included
|
||||
-- in the modified columns)
|
||||
--
|
||||
-- # Example
|
||||
--
|
||||
-- ```sql
|
||||
-- CREATE TABLE users (id SERIAL PRIMARY KEY, updated_at TIMESTAMP NOT NULL DEFAULT NOW());
|
||||
--
|
||||
-- SELECT diesel_manage_updated_at('users');
|
||||
-- ```
|
||||
CREATE OR REPLACE FUNCTION diesel_manage_updated_at(_tbl regclass) RETURNS VOID AS $$
|
||||
BEGIN
|
||||
EXECUTE format('CREATE TRIGGER set_updated_at BEFORE UPDATE ON %s
|
||||
FOR EACH ROW EXECUTE PROCEDURE diesel_set_updated_at()', _tbl);
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
|
||||
CREATE OR REPLACE FUNCTION diesel_set_updated_at() RETURNS trigger AS $$
|
||||
BEGIN
|
||||
IF (
|
||||
NEW IS DISTINCT FROM OLD AND
|
||||
NEW.updated_at IS NOT DISTINCT FROM OLD.updated_at
|
||||
) THEN
|
||||
NEW.updated_at := current_timestamp;
|
||||
END IF;
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ LANGUAGE plpgsql;
|
||||
@@ -1 +0,0 @@
|
||||
DROP TYPE user_role;
|
||||
@@ -1,4 +0,0 @@
|
||||
CREATE TYPE user_role AS ENUM (
|
||||
'STUDENT',
|
||||
'TEACHER',
|
||||
'ADMIN');
|
||||
@@ -1 +0,0 @@
|
||||
DROP TABLE users;
|
||||
@@ -1,11 +0,0 @@
|
||||
CREATE TABLE users
|
||||
(
|
||||
id text PRIMARY KEY NOT NULL,
|
||||
username text UNIQUE NOT NULL,
|
||||
password text NOT NULL,
|
||||
vk_id int4 NULL,
|
||||
access_token text UNIQUE NOT NULL,
|
||||
"group" text NOT NULL,
|
||||
role user_role NOT NULL,
|
||||
version text NOT NULL
|
||||
);
|
||||
@@ -1 +0,0 @@
|
||||
DROP TABLE fcm;
|
||||
@@ -1,6 +0,0 @@
|
||||
CREATE TABLE fcm
|
||||
(
|
||||
user_id text PRIMARY KEY NOT NULL REFERENCES users (id),
|
||||
token text NOT NULL,
|
||||
topics text[] NOT NULL CHECK ( array_position(topics, null) is null )
|
||||
);
|
||||
@@ -1,2 +0,0 @@
|
||||
ALTER TABLE users DROP CONSTRAINT users_telegram_id_key;
|
||||
ALTER TABLE users DROP COLUMN telegram_id;
|
||||
@@ -1,2 +0,0 @@
|
||||
ALTER TABLE users ADD telegram_id int8 NULL;
|
||||
ALTER TABLE users ADD CONSTRAINT users_telegram_id_key UNIQUE (telegram_id);
|
||||
@@ -1,2 +0,0 @@
|
||||
UPDATE users SET "password" = '' WHERE "password" IS NULL;
|
||||
ALTER TABLE users ALTER COLUMN "password" SET NOT NULL;
|
||||
@@ -1 +0,0 @@
|
||||
ALTER TABLE users ALTER COLUMN "password" DROP NOT NULL;
|
||||
@@ -1,3 +0,0 @@
|
||||
UPDATE users SET "android_version" = '' WHERE "android_version" IS NULL;
|
||||
ALTER TABLE users ALTER COLUMN "android_version" SET NOT NULL;
|
||||
ALTER TABLE users RENAME COLUMN android_version TO "version";
|
||||
@@ -1,2 +0,0 @@
|
||||
ALTER TABLE users RENAME COLUMN "version" TO android_version;
|
||||
ALTER TABLE users ALTER COLUMN android_version DROP NOT NULL;
|
||||
@@ -1,2 +0,0 @@
|
||||
UPDATE users SET "group" = '' WHERE "group" IS NULL;
|
||||
ALTER TABLE users ALTER COLUMN "group" SET NOT NULL;
|
||||
@@ -1 +0,0 @@
|
||||
ALTER TABLE users ALTER COLUMN "group" DROP NOT NULL;
|
||||
@@ -1,2 +0,0 @@
|
||||
UPDATE users SET "access_token" = '' WHERE "access_token" IS NULL;
|
||||
ALTER TABLE users ALTER COLUMN "access_token" SET NOT NULL;
|
||||
@@ -1 +0,0 @@
|
||||
ALTER TABLE users ALTER COLUMN "access_token" DROP NOT NULL;
|
||||
@@ -100,6 +100,9 @@ pub enum LessonType {
|
||||
|
||||
/// Защита курсового проекта.
|
||||
CourseProjectDefense,
|
||||
|
||||
/// Практическое занятие.
|
||||
Practice
|
||||
}
|
||||
|
||||
#[derive(Clone, Hash, Debug, Serialize, Deserialize, ToSchema)]
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
[package]
|
||||
name = "provider-engels-polytechnic"
|
||||
version = "0.1.0"
|
||||
version = "0.2.0"
|
||||
edition = "2024"
|
||||
|
||||
[features]
|
||||
@@ -20,7 +20,7 @@ derive_more = { version = "2.0.1", features = ["error", "display"] }
|
||||
|
||||
utoipa = { version = "5.4.0", features = ["macros", "chrono"] }
|
||||
|
||||
calamine = "0.30.0"
|
||||
calamine = "0.30"
|
||||
async-trait = "0.1.89"
|
||||
|
||||
reqwest = "0.12.23"
|
||||
@@ -28,5 +28,5 @@ ua_generator = "0.5.22"
|
||||
regex = "1.11.2"
|
||||
strsim = "0.11.1"
|
||||
log = "0.4.27"
|
||||
sentry = "0.42.0"
|
||||
sentry = "0.43.0"
|
||||
|
||||
|
||||
@@ -25,7 +25,7 @@ pub struct EngelsPolytechnicProvider {
|
||||
}
|
||||
|
||||
impl EngelsPolytechnicProvider {
|
||||
pub async fn new(
|
||||
pub async fn get(
|
||||
update_source: UpdateSource,
|
||||
) -> Result<Arc<dyn ScheduleProvider>, crate::updater::error::Error> {
|
||||
let (updater, snapshot) = Updater::new(update_source).await?;
|
||||
@@ -60,7 +60,7 @@ impl ScheduleProvider for Wrapper {
|
||||
|
||||
log::info!("Updating schedule...");
|
||||
|
||||
match this.updater.update(&mut this.snapshot).await {
|
||||
match this.updater.update(&this.snapshot).await {
|
||||
Ok(snapshot) => {
|
||||
this.snapshot = Arc::new(snapshot);
|
||||
},
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
use crate::or_continue;
|
||||
use crate::parser::error::{ErrorCell, ErrorCellPos};
|
||||
use crate::parser::error::{Error, ErrorCell, ErrorCellPos};
|
||||
use crate::parser::worksheet::WorkSheet;
|
||||
use crate::parser::LessonParseResult::{Lessons, Street};
|
||||
use base::LessonType::Break;
|
||||
@@ -230,7 +230,7 @@ enum LessonParseResult {
|
||||
|
||||
// noinspection GrazieInspection
|
||||
/// Obtaining a non-standard type of lesson by name.
|
||||
fn guess_lesson_type(text: &String) -> Option<LessonType> {
|
||||
fn guess_lesson_type(text: &str) -> Option<LessonType> {
|
||||
static MAP: LazyLock<HashMap<&str, LessonType>> = LazyLock::new(|| {
|
||||
HashMap::from([
|
||||
("консультация", LessonType::Consultation),
|
||||
@@ -240,27 +240,24 @@ fn guess_lesson_type(text: &String) -> Option<LessonType> {
|
||||
("экзамен", LessonType::ExamDefault),
|
||||
("курсовой проект", LessonType::CourseProject),
|
||||
("защита курсового проекта", LessonType::CourseProjectDefense),
|
||||
("практическое занятие", LessonType::Practice),
|
||||
])
|
||||
});
|
||||
|
||||
let name_lower = text.to_lowercase();
|
||||
|
||||
match MAP
|
||||
.iter()
|
||||
.map(|(text, lesson_type)| (lesson_type, strsim::levenshtein(text, &*name_lower)))
|
||||
MAP.iter()
|
||||
.map(|(text, lesson_type)| (lesson_type, strsim::levenshtein(text, &name_lower)))
|
||||
.filter(|x| x.1 <= 4)
|
||||
.min_by_key(|(_, score)| *score)
|
||||
{
|
||||
None => None,
|
||||
Some(v) => Some(v.0.clone()),
|
||||
}
|
||||
.map(|v| v.0.clone())
|
||||
}
|
||||
|
||||
/// Getting a pair or street from a cell.
|
||||
fn parse_lesson(
|
||||
worksheet: &WorkSheet,
|
||||
day: &Day,
|
||||
day_boundaries: &Vec<BoundariesCellInfo>,
|
||||
day_boundaries: &[BoundariesCellInfo],
|
||||
lesson_boundaries: &BoundariesCellInfo,
|
||||
group_column: u32,
|
||||
) -> Result<LessonParseResult, crate::parser::error::Error> {
|
||||
@@ -297,7 +294,7 @@ fn parse_lesson(
|
||||
column: group_column,
|
||||
}))?;
|
||||
|
||||
let range: Option<[u8; 2]> = if lesson_boundaries.default_index != None {
|
||||
let range: Option<[u8; 2]> = if lesson_boundaries.default_index.is_some() {
|
||||
let default = lesson_boundaries.default_index.unwrap() as u8;
|
||||
Some([default, end_time.default_index.unwrap() as u8])
|
||||
} else {
|
||||
@@ -312,7 +309,11 @@ fn parse_lesson(
|
||||
Ok((range, time))
|
||||
}?;
|
||||
|
||||
let (name, mut subgroups, lesson_type) = parse_name_and_subgroups(&name)?;
|
||||
let ParsedLessonName {
|
||||
name,
|
||||
mut subgroups,
|
||||
r#type: lesson_type,
|
||||
} = parse_name_and_subgroups(&name)?;
|
||||
|
||||
{
|
||||
let cabinets: Vec<String> = parse_cabinets(
|
||||
@@ -325,12 +326,10 @@ fn parse_lesson(
|
||||
|
||||
if cab_count == 1 {
|
||||
// Назначаем этот кабинет всем подгруппам
|
||||
let cab = Some(cabinets.get(0).unwrap().clone());
|
||||
let cab = Some(cabinets.first().unwrap().clone());
|
||||
|
||||
for subgroup in &mut subgroups {
|
||||
if let Some(subgroup) = subgroup {
|
||||
subgroup.cabinet = cab.clone()
|
||||
}
|
||||
for subgroup in subgroups.iter_mut().flatten() {
|
||||
subgroup.cabinet = cab.clone()
|
||||
}
|
||||
} else if cab_count == 2 {
|
||||
while subgroups.len() < cab_count {
|
||||
@@ -361,10 +360,7 @@ fn parse_lesson(
|
||||
range: default_range,
|
||||
name: Some(name),
|
||||
time: lesson_time,
|
||||
subgroups: if subgroups.len() == 2
|
||||
&& subgroups.get(0).unwrap().is_none()
|
||||
&& subgroups.get(1).unwrap().is_none()
|
||||
{
|
||||
subgroups: if subgroups.len() == 2 && subgroups.iter().all(|x| x.is_none()) {
|
||||
None
|
||||
} else {
|
||||
Some(subgroups)
|
||||
@@ -416,12 +412,15 @@ fn parse_cabinets(worksheet: &WorkSheet, row_range: (u32, u32), column: u32) ->
|
||||
cabinets
|
||||
}
|
||||
|
||||
struct ParsedLessonName {
|
||||
name: String,
|
||||
subgroups: Vec<Option<LessonSubGroup>>,
|
||||
r#type: Option<LessonType>,
|
||||
}
|
||||
|
||||
//noinspection GrazieInspection
|
||||
/// Getting the "pure" name of the lesson and list of teachers from the text of the lesson cell.
|
||||
fn parse_name_and_subgroups(
|
||||
text: &String,
|
||||
) -> Result<(String, Vec<Option<LessonSubGroup>>, Option<LessonType>), crate::parser::error::Error>
|
||||
{
|
||||
fn parse_name_and_subgroups(text: &str) -> Result<ParsedLessonName, Error> {
|
||||
// Части названия пары:
|
||||
// 1. Само название.
|
||||
// 2. Список преподавателей и подгрупп.
|
||||
@@ -449,7 +448,7 @@ fn parse_name_and_subgroups(
|
||||
|
||||
static NAMES_REGEX: LazyLock<Regex> = LazyLock::new(|| {
|
||||
Regex::new(
|
||||
r"(?:[А-Я][а-я]+\s?(?:[А-Я][\s.]*){2}(?:\(\s*\d\s*[а-я\s]+\))?(?:[\s,]+)?){1,2}+[\s.,]*",
|
||||
r"(?:[А-Я][а-я]+\s?(?:[А-Я][\s.]*){2}(?:\(?\s*\d\s*[а-я\s]+\)?)?(?:[\s,.]+)?){1,2}+[\s.,]*",
|
||||
)
|
||||
.unwrap()
|
||||
});
|
||||
@@ -458,7 +457,7 @@ fn parse_name_and_subgroups(
|
||||
static CLEAN_RE: LazyLock<Regex> = LazyLock::new(|| Regex::new(r"[\s\n\t]+").unwrap());
|
||||
|
||||
let text = CLEAN_RE
|
||||
.replace(&text.replace(&[' ', '\t', '\n'], " "), " ")
|
||||
.replace(&text.replace([' ', '\t', '\n'], " ").replace(",", ""), " ")
|
||||
.to_string();
|
||||
|
||||
let (lesson_name, subgroups, lesson_type) = match NAMES_REGEX.captures(&text) {
|
||||
@@ -466,19 +465,21 @@ fn parse_name_and_subgroups(
|
||||
let capture = captures.get(0).unwrap();
|
||||
|
||||
let subgroups: Vec<Option<LessonSubGroup>> = {
|
||||
let src = capture.as_str().replace(&[' ', '.'], "");
|
||||
let src = capture.as_str().replace([' ', '.'], "");
|
||||
|
||||
let mut shared_subgroup = false;
|
||||
let mut subgroups: [Option<LessonSubGroup>; 2] = [None, None];
|
||||
|
||||
for name in src.split(',') {
|
||||
let open_bracket_index = name.find('(');
|
||||
let digit_index = name.find(|c: char| c.is_ascii_digit());
|
||||
|
||||
let number: u8 = open_bracket_index
|
||||
.map_or(0, |index| name[(index + 1)..(index + 2)].parse().unwrap());
|
||||
let number: u8 =
|
||||
digit_index.map_or(0, |index| name[(index)..(index + 1)].parse().unwrap());
|
||||
|
||||
let teacher_name = {
|
||||
let name_end = open_bracket_index.unwrap_or_else(|| name.len());
|
||||
let name_end = name
|
||||
.find(|c: char| !c.is_alphabetic())
|
||||
.unwrap_or(name.len());
|
||||
|
||||
// Я ебал. Как же я долго до этого доходил.
|
||||
format!(
|
||||
@@ -527,7 +528,7 @@ fn parse_name_and_subgroups(
|
||||
if result.is_none() {
|
||||
#[cfg(not(debug_assertions))]
|
||||
sentry::capture_message(
|
||||
&*format!("Не удалось угадать тип пары '{}'!", extra),
|
||||
&format!("Не удалось угадать тип пары '{}'!", extra),
|
||||
sentry::Level::Warning,
|
||||
);
|
||||
|
||||
@@ -545,7 +546,11 @@ fn parse_name_and_subgroups(
|
||||
None => (text, Vec::new(), None),
|
||||
};
|
||||
|
||||
Ok((lesson_name, subgroups, lesson_type))
|
||||
Ok(ParsedLessonName {
|
||||
name: lesson_name,
|
||||
subgroups,
|
||||
r#type: lesson_type,
|
||||
})
|
||||
}
|
||||
|
||||
/// Getting the start and end of a pair from a cell in the first column of a document.
|
||||
@@ -554,18 +559,11 @@ fn parse_name_and_subgroups(
|
||||
///
|
||||
/// * `cell_data`: text in cell.
|
||||
/// * `date`: date of the current day.
|
||||
fn parse_lesson_boundaries_cell(
|
||||
cell_data: &String,
|
||||
date: DateTime<Utc>,
|
||||
) -> Option<LessonBoundaries> {
|
||||
fn parse_lesson_boundaries_cell(cell_data: &str, date: DateTime<Utc>) -> Option<LessonBoundaries> {
|
||||
static TIME_RE: LazyLock<Regex> =
|
||||
LazyLock::new(|| Regex::new(r"(\d+\.\d+)-(\d+\.\d+)").unwrap());
|
||||
|
||||
let parse_res = if let Some(captures) = TIME_RE.captures(cell_data) {
|
||||
captures
|
||||
} else {
|
||||
return None;
|
||||
};
|
||||
let parse_res = TIME_RE.captures(cell_data)?;
|
||||
|
||||
let start_match = parse_res.get(1).unwrap().as_str();
|
||||
let start_parts: Vec<&str> = start_match.split(".").collect();
|
||||
@@ -579,7 +577,7 @@ fn parse_lesson_boundaries_cell(
|
||||
};
|
||||
|
||||
Some(LessonBoundaries {
|
||||
start: GET_TIME(date.clone(), &start_parts),
|
||||
start: GET_TIME(date, &start_parts),
|
||||
end: GET_TIME(date, &end_parts),
|
||||
})
|
||||
}
|
||||
@@ -607,7 +605,7 @@ fn parse_day_boundaries(
|
||||
continue;
|
||||
};
|
||||
|
||||
let lesson_time = parse_lesson_boundaries_cell(&time_cell, date.clone()).ok_or(
|
||||
let lesson_time = parse_lesson_boundaries_cell(&time_cell, date).ok_or(
|
||||
error::Error::LessonBoundaries(ErrorCell::new(row, column, time_cell.clone())),
|
||||
)?;
|
||||
|
||||
@@ -652,7 +650,7 @@ fn parse_day_boundaries(
|
||||
/// * `week_markup`: markup of the current week.
|
||||
fn parse_week_boundaries(
|
||||
worksheet: &WorkSheet,
|
||||
week_markup: &Vec<DayCellInfo>,
|
||||
week_markup: &[DayCellInfo],
|
||||
) -> Result<Vec<Vec<BoundariesCellInfo>>, crate::parser::error::Error> {
|
||||
let mut result: Vec<Vec<BoundariesCellInfo>> = Vec::new();
|
||||
|
||||
@@ -671,8 +669,8 @@ fn parse_week_boundaries(
|
||||
};
|
||||
|
||||
let day_boundaries = parse_day_boundaries(
|
||||
&worksheet,
|
||||
day_markup.date.clone(),
|
||||
worksheet,
|
||||
day_markup.date,
|
||||
(day_markup.row, end_row),
|
||||
lesson_time_column,
|
||||
)?;
|
||||
@@ -698,7 +696,7 @@ fn convert_groups_to_teachers(
|
||||
.map(|day| Day {
|
||||
name: day.name.clone(),
|
||||
street: day.street.clone(),
|
||||
date: day.date.clone(),
|
||||
date: day.date,
|
||||
lessons: vec![],
|
||||
})
|
||||
.collect();
|
||||
@@ -774,19 +772,6 @@ fn convert_groups_to_teachers(
|
||||
/// * `buffer`: XLS data containing schedule.
|
||||
///
|
||||
/// returns: Result<ParseResult, crate::parser::error::Error>
|
||||
///
|
||||
/// # Examples
|
||||
///
|
||||
/// ```
|
||||
/// use schedule_parser::parse_xls;
|
||||
///
|
||||
/// let result = parse_xls(&include_bytes!("../../schedule.xls").to_vec());
|
||||
///
|
||||
/// assert!(result.is_ok(), "{}", result.err().unwrap());
|
||||
///
|
||||
/// assert_ne!(result.as_ref().unwrap().groups.len(), 0);
|
||||
/// assert_ne!(result.as_ref().unwrap().teachers.len(), 0);
|
||||
/// ```
|
||||
pub fn parse_xls(buffer: &Vec<u8>) -> Result<ParsedSchedule, crate::parser::error::Error> {
|
||||
let cursor = Cursor::new(&buffer);
|
||||
let mut workbook: Xls<_> =
|
||||
@@ -800,7 +785,7 @@ pub fn parse_xls(buffer: &Vec<u8>) -> Result<ParsedSchedule, crate::parser::erro
|
||||
.clone();
|
||||
|
||||
let worksheet_merges = workbook
|
||||
.worksheet_merge_cells(&*worksheet_name)
|
||||
.worksheet_merge_cells(&worksheet_name)
|
||||
.ok_or(error::Error::NoWorkSheets)?;
|
||||
|
||||
WorkSheet {
|
||||
@@ -820,7 +805,7 @@ pub fn parse_xls(buffer: &Vec<u8>) -> Result<ParsedSchedule, crate::parser::erro
|
||||
days: Vec::new(),
|
||||
};
|
||||
|
||||
for day_index in 0..(&week_markup).len() {
|
||||
for day_index in 0..week_markup.len() {
|
||||
let day_markup = &week_markup[day_index];
|
||||
|
||||
let mut day = Day {
|
||||
@@ -836,8 +821,8 @@ pub fn parse_xls(buffer: &Vec<u8>) -> Result<ParsedSchedule, crate::parser::erro
|
||||
match &mut parse_lesson(
|
||||
&worksheet,
|
||||
&day,
|
||||
&day_boundaries,
|
||||
&lesson_boundaries,
|
||||
day_boundaries,
|
||||
lesson_boundaries,
|
||||
group_markup.column,
|
||||
)? {
|
||||
Lessons(lesson) => day.lessons.append(lesson),
|
||||
|
||||
@@ -46,14 +46,17 @@ pub mod error {
|
||||
/// problems with the Yandex Cloud Function invocation.
|
||||
#[display("An error occurred during the request to the Yandex Cloud API: {_0}")]
|
||||
RequestFailed(reqwest::Error),
|
||||
|
||||
#[display("Unable to fetch Uri in 3 retries")]
|
||||
UriFetchFailed,
|
||||
}
|
||||
|
||||
/// Errors that may occur during the creation of a schedule snapshot.
|
||||
#[derive(Debug, Display, Error)]
|
||||
pub enum SnapshotCreationError {
|
||||
/// The URL is the same as the one already being used (no update needed).
|
||||
#[display("The URL is the same as the one already being used.")]
|
||||
SameUrl,
|
||||
/// The ETag is the same (no update needed).
|
||||
#[display("The ETag is the same.")]
|
||||
Same,
|
||||
|
||||
/// The URL query for the XLS file failed to execute, either due to network issues or invalid API parameters.
|
||||
#[display("Failed to fetch URL: {_0}")]
|
||||
@@ -86,11 +89,7 @@ impl Updater {
|
||||
downloader: &mut XlsDownloader,
|
||||
url: String,
|
||||
) -> Result<ScheduleSnapshot, SnapshotCreationError> {
|
||||
if downloader.url.as_ref().is_some_and(|_url| _url.eq(&url)) {
|
||||
return Err(SnapshotCreationError::SameUrl);
|
||||
}
|
||||
|
||||
let head_result = downloader.set_url(&*url).await.map_err(|error| {
|
||||
let head_result = downloader.set_url(&url).await.map_err(|error| {
|
||||
if let FetchError::Unknown(error) = &error {
|
||||
sentry::capture_error(&error);
|
||||
}
|
||||
@@ -98,6 +97,10 @@ impl Updater {
|
||||
SnapshotCreationError::FetchFailed(error)
|
||||
})?;
|
||||
|
||||
if downloader.etag == Some(head_result.etag) {
|
||||
return Err(SnapshotCreationError::Same);
|
||||
}
|
||||
|
||||
let xls_data = downloader
|
||||
.fetch(false)
|
||||
.await
|
||||
@@ -144,18 +147,43 @@ impl Updater {
|
||||
async fn query_url(api_key: &str, func_id: &str) -> Result<String, QueryUrlError> {
|
||||
let client = reqwest::Client::new();
|
||||
|
||||
let uri = client
|
||||
.post(format!(
|
||||
"https://functions.yandexcloud.net/{}?integration=raw",
|
||||
func_id
|
||||
))
|
||||
.header("Authorization", format!("Api-Key {}", api_key))
|
||||
.send()
|
||||
.await
|
||||
.map_err(|error| QueryUrlError::RequestFailed(error))?
|
||||
.text()
|
||||
.await
|
||||
.map_err(|error| QueryUrlError::RequestFailed(error))?;
|
||||
let uri = {
|
||||
// вот бы добавили named-scopes как в котлине,
|
||||
// чтоб мне не пришлось такой хуйнёй страдать.
|
||||
#[allow(unused_assignments)]
|
||||
let mut uri = String::new();
|
||||
let mut counter = 0;
|
||||
|
||||
loop {
|
||||
if counter == 3 {
|
||||
return Err(QueryUrlError::UriFetchFailed);
|
||||
}
|
||||
|
||||
counter += 1;
|
||||
|
||||
uri = client
|
||||
.post(format!(
|
||||
"https://functions.yandexcloud.net/{}?integration=raw",
|
||||
func_id
|
||||
))
|
||||
.header("Authorization", format!("Api-Key {}", api_key))
|
||||
.send()
|
||||
.await
|
||||
.map_err(QueryUrlError::RequestFailed)?
|
||||
.text()
|
||||
.await
|
||||
.map_err(QueryUrlError::RequestFailed)?;
|
||||
|
||||
if uri.is_empty() {
|
||||
log::warn!("[{}] Unable to get uri! Retrying in 5 seconds...", counter);
|
||||
continue;
|
||||
}
|
||||
|
||||
break;
|
||||
}
|
||||
|
||||
uri
|
||||
};
|
||||
|
||||
Ok(format!("https://politehnikum-eng.ru{}", uri.trim()))
|
||||
}
|
||||
@@ -196,7 +224,7 @@ impl Updater {
|
||||
log::info!("Obtaining a link using FaaS...");
|
||||
Self::query_url(yandex_api_key, yandex_func_id)
|
||||
.await
|
||||
.map_err(|error| Error::QueryUrlFailed(error))?
|
||||
.map_err(Error::QueryUrlFailed)?
|
||||
}
|
||||
_ => unreachable!(),
|
||||
};
|
||||
@@ -205,7 +233,7 @@ impl Updater {
|
||||
|
||||
let snapshot = Self::new_snapshot(&mut this.downloader, url)
|
||||
.await
|
||||
.map_err(|error| Error::SnapshotCreationFailed(error))?;
|
||||
.map_err(Error::SnapshotCreationFailed)?;
|
||||
|
||||
log::info!("Schedule snapshot successfully created!");
|
||||
|
||||
@@ -243,13 +271,13 @@ impl Updater {
|
||||
yandex_func_id,
|
||||
} => Self::query_url(yandex_api_key.as_str(), yandex_func_id.as_str())
|
||||
.await
|
||||
.map_err(|error| Error::QueryUrlFailed(error))?,
|
||||
.map_err(Error::QueryUrlFailed)?,
|
||||
_ => unreachable!(),
|
||||
};
|
||||
|
||||
let snapshot = match Self::new_snapshot(&mut self.downloader, url).await {
|
||||
Ok(snapshot) => snapshot,
|
||||
Err(SnapshotCreationError::SameUrl) => {
|
||||
Err(SnapshotCreationError::Same) => {
|
||||
let mut clone = current_snapshot.clone();
|
||||
clone.update();
|
||||
|
||||
|
||||
@@ -66,25 +66,30 @@ pub struct FetchOk {
|
||||
/// Date data received.
|
||||
pub requested_at: DateTime<Utc>,
|
||||
|
||||
/// Etag.
|
||||
pub etag: String,
|
||||
|
||||
/// File data.
|
||||
pub data: Option<Vec<u8>>,
|
||||
}
|
||||
|
||||
impl FetchOk {
|
||||
/// Result without file content.
|
||||
pub fn head(uploaded_at: DateTime<Utc>) -> Self {
|
||||
pub fn head(uploaded_at: DateTime<Utc>, etag: String) -> Self {
|
||||
FetchOk {
|
||||
uploaded_at,
|
||||
requested_at: Utc::now(),
|
||||
etag,
|
||||
data: None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Full result.
|
||||
pub fn get(uploaded_at: DateTime<Utc>, data: Vec<u8>) -> Self {
|
||||
pub fn get(uploaded_at: DateTime<Utc>, etag: String, data: Vec<u8>) -> Self {
|
||||
FetchOk {
|
||||
uploaded_at,
|
||||
requested_at: Utc::now(),
|
||||
etag,
|
||||
data: Some(data),
|
||||
}
|
||||
}
|
||||
@@ -94,11 +99,15 @@ pub type FetchResult = Result<FetchOk, FetchError>;
|
||||
|
||||
pub struct XlsDownloader {
|
||||
pub url: Option<String>,
|
||||
pub etag: Option<String>,
|
||||
}
|
||||
|
||||
impl XlsDownloader {
|
||||
pub fn new() -> Self {
|
||||
XlsDownloader { url: None }
|
||||
XlsDownloader {
|
||||
url: None,
|
||||
etag: None,
|
||||
}
|
||||
}
|
||||
|
||||
async fn fetch_specified(url: &str, head: bool) -> FetchResult {
|
||||
@@ -124,9 +133,12 @@ impl XlsDownloader {
|
||||
.get("Content-Type")
|
||||
.ok_or(FetchError::bad_headers("Content-Type"))?;
|
||||
|
||||
if !headers.contains_key("etag") {
|
||||
return Err(FetchError::bad_headers("etag"));
|
||||
}
|
||||
let etag = headers
|
||||
.get("etag")
|
||||
.ok_or(FetchError::bad_headers("etag"))?
|
||||
.to_str()
|
||||
.or(Err(FetchError::bad_headers("etag")))?
|
||||
.to_string();
|
||||
|
||||
let last_modified = headers
|
||||
.get("last-modified")
|
||||
@@ -136,14 +148,18 @@ impl XlsDownloader {
|
||||
return Err(FetchError::bad_content_type(content_type.to_str().unwrap()));
|
||||
}
|
||||
|
||||
let last_modified = DateTime::parse_from_rfc2822(&last_modified.to_str().unwrap())
|
||||
let last_modified = DateTime::parse_from_rfc2822(last_modified.to_str().unwrap())
|
||||
.unwrap()
|
||||
.with_timezone(&Utc);
|
||||
|
||||
Ok(if head {
|
||||
FetchOk::head(last_modified)
|
||||
FetchOk::head(last_modified, etag)
|
||||
} else {
|
||||
FetchOk::get(last_modified, response.bytes().await.unwrap().to_vec())
|
||||
FetchOk::get(
|
||||
last_modified,
|
||||
etag,
|
||||
response.bytes().await.unwrap().to_vec(),
|
||||
)
|
||||
})
|
||||
}
|
||||
|
||||
@@ -151,14 +167,14 @@ impl XlsDownloader {
|
||||
if self.url.is_none() {
|
||||
Err(FetchError::NoUrlProvided)
|
||||
} else {
|
||||
Self::fetch_specified(&*self.url.as_ref().unwrap(), head).await
|
||||
Self::fetch_specified(self.url.as_ref().unwrap(), head).await
|
||||
}
|
||||
}
|
||||
|
||||
pub async fn set_url(&mut self, url: &str) -> FetchResult {
|
||||
let result = Self::fetch_specified(url, true).await;
|
||||
|
||||
if let Ok(_) = result {
|
||||
if result.is_ok() {
|
||||
self.url = Some(url.to_string());
|
||||
}
|
||||
|
||||
|
||||
@@ -1,148 +0,0 @@
|
||||
pub mod users {
|
||||
use crate::database::models::User;
|
||||
use crate::database::schema::users::dsl::users;
|
||||
use crate::database::schema::users::dsl::*;
|
||||
use crate::state::AppState;
|
||||
use actix_web::web;
|
||||
use diesel::{insert_into, ExpressionMethods, QueryResult};
|
||||
use diesel::{QueryDsl, RunQueryDsl};
|
||||
use diesel::{SaveChangesDsl, SelectableHelper};
|
||||
use std::ops::DerefMut;
|
||||
|
||||
pub async fn get(state: &web::Data<AppState>, _id: &String) -> QueryResult<User> {
|
||||
users
|
||||
.filter(id.eq(_id))
|
||||
.select(User::as_select())
|
||||
.first(state.get_database().await.deref_mut())
|
||||
}
|
||||
|
||||
pub async fn get_by_username(
|
||||
state: &web::Data<AppState>,
|
||||
_username: &String,
|
||||
) -> QueryResult<User> {
|
||||
users
|
||||
.filter(username.eq(_username))
|
||||
.select(User::as_select())
|
||||
.first(state.get_database().await.deref_mut())
|
||||
}
|
||||
|
||||
//noinspection RsTraitObligations
|
||||
pub async fn get_by_vk_id(state: &web::Data<AppState>, _vk_id: i32) -> QueryResult<User> {
|
||||
users
|
||||
.filter(vk_id.eq(_vk_id))
|
||||
.select(User::as_select())
|
||||
.first(state.get_database().await.deref_mut())
|
||||
}
|
||||
|
||||
//noinspection RsTraitObligations
|
||||
pub async fn get_by_telegram_id(
|
||||
state: &web::Data<AppState>,
|
||||
_telegram_id: i64,
|
||||
) -> QueryResult<User> {
|
||||
users
|
||||
.filter(telegram_id.eq(_telegram_id))
|
||||
.select(User::as_select())
|
||||
.first(state.get_database().await.deref_mut())
|
||||
}
|
||||
|
||||
//noinspection DuplicatedCode
|
||||
pub async fn contains_by_username(state: &web::Data<AppState>, _username: &String) -> bool {
|
||||
// и как это нахуй сократить блять примеров нихуя нет, нихуя не работает
|
||||
// как меня этот раст заебал уже
|
||||
|
||||
match users
|
||||
.filter(username.eq(_username))
|
||||
.count()
|
||||
.get_result::<i64>(state.get_database().await.deref_mut())
|
||||
{
|
||||
Ok(count) => count > 0,
|
||||
Err(_) => false,
|
||||
}
|
||||
}
|
||||
|
||||
//noinspection DuplicatedCode
|
||||
//noinspection RsTraitObligations
|
||||
pub async fn contains_by_vk_id(state: &web::Data<AppState>, _vk_id: i32) -> bool {
|
||||
match users
|
||||
.filter(vk_id.eq(_vk_id))
|
||||
.count()
|
||||
.get_result::<i64>(state.get_database().await.deref_mut())
|
||||
{
|
||||
Ok(count) => count > 0,
|
||||
Err(_) => false,
|
||||
}
|
||||
}
|
||||
|
||||
pub async fn insert(state: &web::Data<AppState>, user: &User) -> QueryResult<usize> {
|
||||
insert_into(users)
|
||||
.values(user)
|
||||
.execute(state.get_database().await.deref_mut())
|
||||
}
|
||||
|
||||
/// Function declaration [User::save][UserSave::save].
|
||||
pub trait UserSave {
|
||||
/// Saves the user's changes to the database.
|
||||
///
|
||||
/// # Arguments
|
||||
///
|
||||
/// * `state`: The state of the actix-web application that stores the mutex of the [connection][diesel::PgConnection].
|
||||
///
|
||||
/// returns: `QueryResult<User>`
|
||||
///
|
||||
/// # Examples
|
||||
///
|
||||
/// ```
|
||||
/// use crate::database::driver::users;
|
||||
///
|
||||
/// #[derive(Deserialize)]
|
||||
/// struct Params {
|
||||
/// pub username: String,
|
||||
/// }
|
||||
///
|
||||
/// #[patch("/")]
|
||||
/// async fn patch_user(
|
||||
/// app_state: web::Data<AppState>,
|
||||
/// user: SyncExtractor<User>,
|
||||
/// web::Query(params): web::Query<Params>,
|
||||
/// ) -> web::Json<User> {
|
||||
/// let mut user = user.into_inner();
|
||||
///
|
||||
/// user.username = params.username;
|
||||
///
|
||||
/// match user.save(&app_state) {
|
||||
/// Ok(user) => web::Json(user),
|
||||
/// Err(e) => {
|
||||
/// eprintln!("Failed to save user: {e}");
|
||||
/// panic!();
|
||||
/// }
|
||||
/// }
|
||||
/// }
|
||||
/// ```
|
||||
async fn save(&self, state: &web::Data<AppState>) -> QueryResult<User>;
|
||||
}
|
||||
|
||||
/// Implementation of [UserSave][UserSave] trait.
|
||||
impl UserSave for User {
|
||||
async fn save(&self, state: &web::Data<AppState>) -> QueryResult<User> {
|
||||
self.save_changes::<Self>(state.get_database().await.deref_mut())
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
pub async fn delete_by_username(state: &web::Data<AppState>, _username: &String) -> bool {
|
||||
match diesel::delete(users.filter(username.eq(_username)))
|
||||
.execute(state.get_database().await.deref_mut())
|
||||
{
|
||||
Ok(count) => count > 0,
|
||||
Err(_) => false,
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
pub async fn insert_or_ignore(state: &web::Data<AppState>, user: &User) -> QueryResult<usize> {
|
||||
insert_into(users)
|
||||
.values(user)
|
||||
.on_conflict_do_nothing()
|
||||
.execute(state.get_database().await.deref_mut())
|
||||
}
|
||||
}
|
||||
@@ -1,3 +0,0 @@
|
||||
pub mod driver;
|
||||
pub mod models;
|
||||
pub mod schema;
|
||||
@@ -1,87 +0,0 @@
|
||||
use actix_macros::ResponderJson;
|
||||
use diesel::QueryId;
|
||||
use diesel::prelude::*;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use utoipa::ToSchema;
|
||||
|
||||
#[derive(
|
||||
Copy, Clone, PartialEq, Debug, Serialize, Deserialize, diesel_derive_enum::DbEnum, ToSchema,
|
||||
)]
|
||||
#[ExistingTypePath = "crate::database::schema::sql_types::UserRole"]
|
||||
#[DbValueStyle = "UPPERCASE"]
|
||||
#[serde(rename_all = "UPPERCASE")]
|
||||
pub enum UserRole {
|
||||
Student,
|
||||
Teacher,
|
||||
Admin,
|
||||
}
|
||||
|
||||
#[derive(
|
||||
Identifiable,
|
||||
AsChangeset,
|
||||
Queryable,
|
||||
QueryId,
|
||||
Selectable,
|
||||
Serialize,
|
||||
Insertable,
|
||||
Debug,
|
||||
ToSchema,
|
||||
ResponderJson,
|
||||
)]
|
||||
#[diesel(table_name = crate::database::schema::users)]
|
||||
#[diesel(treat_none_as_null = true)]
|
||||
pub struct User {
|
||||
/// Account UUID.
|
||||
pub id: String,
|
||||
|
||||
/// User name.
|
||||
pub username: String,
|
||||
|
||||
/// BCrypt password hash.
|
||||
pub password: Option<String>,
|
||||
|
||||
/// ID of the linked VK account.
|
||||
pub vk_id: Option<i32>,
|
||||
|
||||
/// JWT access token.
|
||||
pub access_token: Option<String>,
|
||||
|
||||
/// Group.
|
||||
pub group: Option<String>,
|
||||
|
||||
/// Role.
|
||||
pub role: UserRole,
|
||||
|
||||
/// Version of the installed Polytechnic+ application.
|
||||
pub android_version: Option<String>,
|
||||
|
||||
/// ID of the linked Telegram account.
|
||||
pub telegram_id: Option<i64>,
|
||||
}
|
||||
|
||||
#[derive(
|
||||
Debug,
|
||||
Clone,
|
||||
Serialize,
|
||||
Identifiable,
|
||||
Queryable,
|
||||
Selectable,
|
||||
Insertable,
|
||||
AsChangeset,
|
||||
Associations,
|
||||
ToSchema,
|
||||
ResponderJson,
|
||||
)]
|
||||
#[diesel(belongs_to(User))]
|
||||
#[diesel(table_name = crate::database::schema::fcm)]
|
||||
#[diesel(primary_key(user_id))]
|
||||
pub struct FCM {
|
||||
/// Account UUID.
|
||||
pub user_id: String,
|
||||
|
||||
/// FCM token.
|
||||
pub token: String,
|
||||
|
||||
/// List of topics subscribed to by the user.
|
||||
pub topics: Vec<Option<String>>,
|
||||
}
|
||||
@@ -1,39 +0,0 @@
|
||||
// @generated automatically by Diesel CLI.
|
||||
|
||||
pub mod sql_types {
|
||||
#[derive(diesel::query_builder::QueryId, Clone, diesel::sql_types::SqlType)]
|
||||
#[diesel(postgres_type(name = "user_role"))]
|
||||
pub struct UserRole;
|
||||
}
|
||||
|
||||
diesel::table! {
|
||||
fcm (user_id) {
|
||||
user_id -> Text,
|
||||
token -> Text,
|
||||
topics -> Array<Nullable<Text>>,
|
||||
}
|
||||
}
|
||||
|
||||
diesel::table! {
|
||||
use diesel::sql_types::*;
|
||||
use super::sql_types::UserRole;
|
||||
|
||||
users (id) {
|
||||
id -> Text,
|
||||
username -> Text,
|
||||
password -> Nullable<Text>,
|
||||
vk_id -> Nullable<Int4>,
|
||||
access_token -> Nullable<Text>,
|
||||
group -> Nullable<Text>,
|
||||
role -> UserRole,
|
||||
android_version -> Nullable<Text>,
|
||||
telegram_id -> Nullable<Int8>,
|
||||
}
|
||||
}
|
||||
|
||||
diesel::joinable!(fcm -> users (user_id));
|
||||
|
||||
diesel::allow_tables_to_appear_in_same_query!(
|
||||
fcm,
|
||||
users,
|
||||
);
|
||||
@@ -1,5 +1,3 @@
|
||||
use crate::database::driver;
|
||||
use crate::database::models::User;
|
||||
use crate::extractors::base::FromRequestAsync;
|
||||
use crate::state::AppState;
|
||||
use crate::utility::jwt;
|
||||
@@ -8,6 +6,8 @@ use actix_web::body::BoxBody;
|
||||
use actix_web::dev::Payload;
|
||||
use actix_web::http::header;
|
||||
use actix_web::{web, HttpRequest};
|
||||
use database::entity::User;
|
||||
use database::query::Query;
|
||||
use derive_more::Display;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use std::fmt::Debug;
|
||||
@@ -88,10 +88,20 @@ impl FromRequestAsync for User {
|
||||
let user_id = jwt::verify_and_decode(&access_token)
|
||||
.map_err(|_| Error::InvalidAccessToken.into_err())?;
|
||||
|
||||
let app_state = req.app_data::<web::Data<AppState>>().unwrap();
|
||||
let db = req
|
||||
.app_data::<web::Data<AppState>>()
|
||||
.unwrap()
|
||||
.get_database();
|
||||
|
||||
driver::users::get(app_state, &user_id)
|
||||
Query::find_user_by_id(db, &user_id)
|
||||
.await
|
||||
.map_err(|_| Error::NoUser.into())
|
||||
.and_then(|user| {
|
||||
if let Some(user) = user {
|
||||
Ok(user)
|
||||
} else {
|
||||
Err(actix_web::Error::from(Error::NoUser))
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
@@ -5,7 +5,6 @@ use std::future::{Ready, ready};
|
||||
use std::ops;
|
||||
|
||||
/// # Async extractor.
|
||||
|
||||
/// Asynchronous object extractor from a query.
|
||||
pub struct AsyncExtractor<T>(T);
|
||||
|
||||
@@ -80,7 +79,6 @@ impl<T: FromRequestAsync> FromRequest for AsyncExtractor<T> {
|
||||
}
|
||||
|
||||
/// # Sync extractor.
|
||||
|
||||
/// Synchronous object extractor from a query.
|
||||
pub struct SyncExtractor<T>(T);
|
||||
|
||||
|
||||
@@ -12,8 +12,6 @@ use utoipa_rapidoc::RapiDoc;
|
||||
|
||||
mod state;
|
||||
|
||||
mod database;
|
||||
|
||||
mod extractors;
|
||||
mod middlewares;
|
||||
mod routes;
|
||||
@@ -72,7 +70,7 @@ pub fn get_api_scope<
|
||||
async fn async_main() -> io::Result<()> {
|
||||
info!("Запуск сервера...");
|
||||
|
||||
let app_state = new_app_state().await.unwrap();
|
||||
let app_state = new_app_state(None).await.unwrap();
|
||||
|
||||
HttpServer::new(move || {
|
||||
let (app, api) = App::new()
|
||||
@@ -111,7 +109,7 @@ fn main() -> io::Result<()> {
|
||||
},
|
||||
));
|
||||
|
||||
dotenv().unwrap();
|
||||
let _ = dotenv();
|
||||
|
||||
env_logger::init();
|
||||
|
||||
|
||||
@@ -1,25 +1,20 @@
|
||||
use crate::database::models::User;
|
||||
use crate::extractors::authorized_user;
|
||||
use crate::extractors::base::FromRequestAsync;
|
||||
use actix_web::body::{BoxBody, EitherBody};
|
||||
use actix_web::dev::{Payload, Service, ServiceRequest, ServiceResponse, Transform, forward_ready};
|
||||
use actix_web::dev::{forward_ready, Payload, Service, ServiceRequest, ServiceResponse, Transform};
|
||||
use actix_web::{Error, HttpRequest, ResponseError};
|
||||
use database::entity::User;
|
||||
use futures_util::future::LocalBoxFuture;
|
||||
use std::future::{Ready, ready};
|
||||
use std::future::{ready, Ready};
|
||||
use std::rc::Rc;
|
||||
|
||||
/// Middleware guard working with JWT tokens.
|
||||
#[derive(Default)]
|
||||
pub struct JWTAuthorization {
|
||||
/// List of ignored endpoints.
|
||||
pub ignore: &'static [&'static str],
|
||||
}
|
||||
|
||||
impl Default for JWTAuthorization {
|
||||
fn default() -> Self {
|
||||
Self { ignore: &[] }
|
||||
}
|
||||
}
|
||||
|
||||
impl<S, B> Transform<S, ServiceRequest> for JWTAuthorization
|
||||
where
|
||||
S: Service<ServiceRequest, Response = ServiceResponse<B>, Error = Error> + 'static,
|
||||
@@ -70,8 +65,8 @@ where
|
||||
return false;
|
||||
}
|
||||
|
||||
if let Some(other) = path.as_bytes().iter().nth(ignore.len()) {
|
||||
return ['?' as u8, '/' as u8].contains(other);
|
||||
if let Some(other) = path.as_bytes().get(ignore.len()) {
|
||||
return [b'?', b'/'].contains(other);
|
||||
}
|
||||
|
||||
true
|
||||
|
||||
@@ -1,10 +1,10 @@
|
||||
use actix_web::Error;
|
||||
use actix_web::body::{BoxBody, EitherBody};
|
||||
use actix_web::dev::{Service, ServiceRequest, ServiceResponse, Transform, forward_ready};
|
||||
use actix_web::dev::{forward_ready, Service, ServiceRequest, ServiceResponse, Transform};
|
||||
use actix_web::http::header;
|
||||
use actix_web::http::header::HeaderValue;
|
||||
use actix_web::Error;
|
||||
use futures_util::future::LocalBoxFuture;
|
||||
use std::future::{Ready, ready};
|
||||
use std::future::{ready, Ready};
|
||||
|
||||
/// Middleware to specify the encoding in the Content-Type header.
|
||||
pub struct ContentTypeBootstrap;
|
||||
@@ -30,7 +30,7 @@ pub struct ContentTypeMiddleware<S> {
|
||||
service: S,
|
||||
}
|
||||
|
||||
impl<'a, S, B> Service<ServiceRequest> for ContentTypeMiddleware<S>
|
||||
impl<S, B> Service<ServiceRequest> for ContentTypeMiddleware<S>
|
||||
where
|
||||
S: Service<ServiceRequest, Response = ServiceResponse<B>, Error = Error>,
|
||||
S::Future: 'static,
|
||||
@@ -49,13 +49,14 @@ where
|
||||
let mut response = fut.await?;
|
||||
|
||||
let headers = response.response_mut().headers_mut();
|
||||
if let Some(content_type) = headers.get("Content-Type") {
|
||||
if content_type == "application/json" {
|
||||
headers.insert(
|
||||
header::CONTENT_TYPE,
|
||||
HeaderValue::from_static("application/json; charset=utf8"),
|
||||
);
|
||||
}
|
||||
|
||||
if let Some(content_type) = headers.get("Content-Type")
|
||||
&& content_type == "application/json"
|
||||
{
|
||||
headers.insert(
|
||||
header::CONTENT_TYPE,
|
||||
HeaderValue::from_static("application/json; charset=utf8"),
|
||||
);
|
||||
}
|
||||
|
||||
Ok(response.map_into_left_body())
|
||||
|
||||
@@ -2,16 +2,6 @@ use jsonwebtoken::errors::ErrorKind;
|
||||
use jsonwebtoken::{Algorithm, DecodingKey, Validation, decode};
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
#[derive(Deserialize, Serialize)]
|
||||
struct TokenData {
|
||||
iis: String,
|
||||
sub: i32,
|
||||
app: i32,
|
||||
exp: i32,
|
||||
iat: i32,
|
||||
jti: i32,
|
||||
}
|
||||
|
||||
#[derive(Debug, Serialize, Deserialize)]
|
||||
struct Claims {
|
||||
sub: i32,
|
||||
@@ -22,7 +12,7 @@ struct Claims {
|
||||
|
||||
#[derive(Debug, PartialEq)]
|
||||
pub enum Error {
|
||||
JwtError(ErrorKind),
|
||||
Jwt(ErrorKind),
|
||||
InvalidSignature,
|
||||
InvalidToken,
|
||||
Expired,
|
||||
@@ -49,10 +39,10 @@ const VK_PUBLIC_KEY: &str = concat!(
|
||||
"-----END PUBLIC KEY-----"
|
||||
);
|
||||
|
||||
pub fn parse_vk_id(token_str: &String, client_id: i32) -> Result<i32, Error> {
|
||||
pub fn parse_vk_id(token_str: &str, client_id: i32) -> Result<i32, Error> {
|
||||
let dkey = DecodingKey::from_rsa_pem(VK_PUBLIC_KEY.as_bytes()).unwrap();
|
||||
|
||||
match decode::<Claims>(&token_str, &dkey, &Validation::new(Algorithm::RS256)) {
|
||||
match decode::<Claims>(token_str, &dkey, &Validation::new(Algorithm::RS256)) {
|
||||
Ok(token_data) => {
|
||||
let claims = token_data.claims;
|
||||
|
||||
@@ -77,7 +67,7 @@ pub fn parse_vk_id(token_str: &String, client_id: i32) -> Result<i32, Error> {
|
||||
ErrorKind::Base64(_) => Error::InvalidToken,
|
||||
ErrorKind::Json(_) => Error::InvalidToken,
|
||||
ErrorKind::Utf8(_) => Error::InvalidToken,
|
||||
kind => Error::JwtError(kind),
|
||||
kind => Error::Jwt(kind),
|
||||
}),
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,31 +1,34 @@
|
||||
use self::schema::*;
|
||||
use crate::database::driver;
|
||||
use crate::database::driver::users::UserSave;
|
||||
use crate::routes::auth::shared::parse_vk_id;
|
||||
use crate::routes::auth::sign_in::schema::SignInData::{Default, VkOAuth};
|
||||
use crate::routes::schema::ResponseError;
|
||||
use crate::routes::schema::user::UserResponse;
|
||||
use crate::{AppState, utility};
|
||||
use crate::routes::schema::ResponseError;
|
||||
use crate::{utility, AppState};
|
||||
use actix_web::{post, web};
|
||||
use database::query::Query;
|
||||
use web::Json;
|
||||
|
||||
async fn sign_in_combined(
|
||||
data: SignInData,
|
||||
app_state: &web::Data<AppState>,
|
||||
) -> Result<UserResponse, ErrorCode> {
|
||||
let db = app_state.get_database();
|
||||
|
||||
let user = match &data {
|
||||
Default(data) => driver::users::get_by_username(&app_state, &data.username).await,
|
||||
VkOAuth(id) => driver::users::get_by_vk_id(&app_state, *id).await,
|
||||
};
|
||||
Default(data) => Query::find_user_by_username(db, &data.username).await,
|
||||
VkOAuth(id) => Query::find_user_by_vk_id(db, *id).await,
|
||||
}
|
||||
.ok()
|
||||
.flatten();
|
||||
|
||||
match user {
|
||||
Ok(mut user) => {
|
||||
Some(user) => {
|
||||
if let Default(data) = data {
|
||||
if user.password.is_none() {
|
||||
return Err(ErrorCode::IncorrectCredentials);
|
||||
}
|
||||
|
||||
match bcrypt::verify(&data.password, &user.password.as_ref().unwrap()) {
|
||||
match bcrypt::verify(&data.password, user.password.as_ref().unwrap()) {
|
||||
Ok(result) => {
|
||||
if !result {
|
||||
return Err(ErrorCode::IncorrectCredentials);
|
||||
@@ -37,14 +40,11 @@ async fn sign_in_combined(
|
||||
}
|
||||
}
|
||||
|
||||
user.access_token = Some(utility::jwt::encode(&user.id));
|
||||
|
||||
user.save(&app_state).await.expect("Failed to update user");
|
||||
|
||||
Ok(user.into())
|
||||
let access_token = utility::jwt::encode(&user.id);
|
||||
Ok(UserResponse::from_user_with_token(user, access_token))
|
||||
}
|
||||
|
||||
Err(_) => Err(ErrorCode::IncorrectCredentials),
|
||||
None => Err(ErrorCode::IncorrectCredentials),
|
||||
}
|
||||
}
|
||||
|
||||
@@ -124,8 +124,6 @@ mod schema {
|
||||
InvalidVkAccessToken,
|
||||
}
|
||||
|
||||
/// Internal
|
||||
|
||||
/// Type of authorization.
|
||||
pub enum SignInData {
|
||||
/// User and password name and password.
|
||||
@@ -139,16 +137,16 @@ mod schema {
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::schema::*;
|
||||
use crate::database::driver;
|
||||
use crate::database::models::{User, UserRole};
|
||||
use crate::routes::auth::sign_in::sign_in;
|
||||
use crate::test_env::tests::{static_app_state, test_app_state, test_env};
|
||||
use crate::utility;
|
||||
use actix_test::test_app;
|
||||
use actix_web::dev::ServiceResponse;
|
||||
use actix_web::http::Method;
|
||||
use actix_web::http::StatusCode;
|
||||
use actix_web::test;
|
||||
use database::entity::sea_orm_active_enums::UserRole;
|
||||
use database::entity::ActiveUser;
|
||||
use database::sea_orm::{ActiveModelTrait, Set};
|
||||
use sha1::{Digest, Sha1};
|
||||
use std::fmt::Write;
|
||||
|
||||
@@ -182,22 +180,24 @@ mod tests {
|
||||
test_env();
|
||||
|
||||
let app_state = static_app_state().await;
|
||||
driver::users::insert_or_ignore(
|
||||
&app_state,
|
||||
&User {
|
||||
id: id.clone(),
|
||||
username,
|
||||
password: Some(bcrypt::hash("example".to_string(), bcrypt::DEFAULT_COST).unwrap()),
|
||||
vk_id: None,
|
||||
telegram_id: None,
|
||||
access_token: Some(utility::jwt::encode(&id)),
|
||||
group: Some("ИС-214/23".to_string()),
|
||||
role: UserRole::Student,
|
||||
android_version: None,
|
||||
},
|
||||
)
|
||||
.await
|
||||
.unwrap();
|
||||
|
||||
let active_user = ActiveUser {
|
||||
id: Set(id.clone()),
|
||||
username: Set(username),
|
||||
password: Set(Some(
|
||||
bcrypt::hash("example", bcrypt::DEFAULT_COST).unwrap(),
|
||||
)),
|
||||
vk_id: Set(None),
|
||||
telegram_id: Set(None),
|
||||
group: Set(Some("ИС-214/23".to_string())),
|
||||
role: Set(UserRole::Student),
|
||||
android_version: Set(None),
|
||||
};
|
||||
|
||||
active_user
|
||||
.save(app_state.get_database())
|
||||
.await
|
||||
.expect("Failed to save user");
|
||||
}
|
||||
|
||||
#[actix_web::test]
|
||||
|
||||
@@ -1,11 +1,13 @@
|
||||
use self::schema::*;
|
||||
use crate::AppState;
|
||||
use crate::database::driver;
|
||||
use crate::database::models::UserRole;
|
||||
use crate::routes::auth::shared::parse_vk_id;
|
||||
use crate::routes::schema::ResponseError;
|
||||
use crate::routes::schema::user::UserResponse;
|
||||
use crate::routes::schema::ResponseError;
|
||||
use crate::{utility, AppState};
|
||||
use actix_web::{post, web};
|
||||
use database::entity::sea_orm_active_enums::UserRole;
|
||||
use database::entity::ActiveUser;
|
||||
use database::query::Query;
|
||||
use database::sea_orm::ActiveModelTrait;
|
||||
use web::Json;
|
||||
|
||||
async fn sign_up_combined(
|
||||
@@ -28,22 +30,30 @@ async fn sign_up_combined(
|
||||
return Err(ErrorCode::InvalidGroupName);
|
||||
}
|
||||
|
||||
// If user with specified username already exists.
|
||||
if driver::users::contains_by_username(&app_state, &data.username).await {
|
||||
let db = app_state.get_database();
|
||||
|
||||
// If user with specified username already exists.O
|
||||
if Query::find_user_by_username(db, &data.username)
|
||||
.await
|
||||
.is_ok_and(|user| user.is_some())
|
||||
{
|
||||
return Err(ErrorCode::UsernameAlreadyExists);
|
||||
}
|
||||
|
||||
// If user with specified VKID already exists.
|
||||
if let Some(id) = data.vk_id {
|
||||
if driver::users::contains_by_vk_id(&app_state, id).await {
|
||||
return Err(ErrorCode::VkAlreadyExists);
|
||||
}
|
||||
if let Some(id) = data.vk_id
|
||||
&& Query::is_user_exists_by_vk_id(db, id)
|
||||
.await
|
||||
.expect("Failed to check user existence")
|
||||
{
|
||||
return Err(ErrorCode::VkAlreadyExists);
|
||||
}
|
||||
|
||||
let user = data.into();
|
||||
driver::users::insert(&app_state, &user).await.unwrap();
|
||||
let active_user: ActiveUser = data.into();
|
||||
let user = active_user.insert(db).await.unwrap();
|
||||
let access_token = utility::jwt::encode(&user.id);
|
||||
|
||||
Ok(UserResponse::from(&user)).into()
|
||||
Ok(UserResponse::from_user_with_token(user, access_token))
|
||||
}
|
||||
|
||||
#[utoipa::path(responses(
|
||||
@@ -101,10 +111,11 @@ pub async fn sign_up_vk(
|
||||
}
|
||||
|
||||
mod schema {
|
||||
use crate::database::models::{User, UserRole};
|
||||
use crate::routes::schema::user::UserResponse;
|
||||
use crate::utility;
|
||||
use actix_macros::ErrResponse;
|
||||
use database::entity::sea_orm_active_enums::UserRole;
|
||||
use database::entity::ActiveUser;
|
||||
use database::sea_orm::Set;
|
||||
use derive_more::Display;
|
||||
use objectid::ObjectId;
|
||||
use serde::{Deserialize, Serialize};
|
||||
@@ -134,7 +145,7 @@ mod schema {
|
||||
}
|
||||
|
||||
pub mod vk {
|
||||
use crate::database::models::UserRole;
|
||||
use database::entity::sea_orm_active_enums::UserRole;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
#[derive(Serialize, Deserialize, utoipa::ToSchema)]
|
||||
@@ -189,8 +200,6 @@ mod schema {
|
||||
VkAlreadyExists,
|
||||
}
|
||||
|
||||
/// Internal
|
||||
|
||||
/// Data for registration.
|
||||
pub struct SignUpData {
|
||||
// TODO: сделать ограничение на минимальную и максимальную длину при регистрации и смене.
|
||||
@@ -215,25 +224,21 @@ mod schema {
|
||||
pub version: String,
|
||||
}
|
||||
|
||||
impl Into<User> for SignUpData {
|
||||
fn into(self) -> User {
|
||||
assert_ne!(self.password.is_some(), self.vk_id.is_some());
|
||||
impl From<SignUpData> for ActiveUser {
|
||||
fn from(value: SignUpData) -> Self {
|
||||
assert_ne!(value.password.is_some(), value.vk_id.is_some());
|
||||
|
||||
let id = ObjectId::new().unwrap().to_string();
|
||||
let access_token = Some(utility::jwt::encode(&id));
|
||||
|
||||
User {
|
||||
id,
|
||||
username: self.username,
|
||||
password: self
|
||||
ActiveUser {
|
||||
id: Set(ObjectId::new().unwrap().to_string()),
|
||||
username: Set(value.username),
|
||||
password: Set(value
|
||||
.password
|
||||
.map(|x| bcrypt::hash(x, bcrypt::DEFAULT_COST).unwrap()),
|
||||
vk_id: self.vk_id,
|
||||
telegram_id: None,
|
||||
access_token,
|
||||
group: Some(self.group),
|
||||
role: self.role,
|
||||
android_version: Some(self.version),
|
||||
.map(|x| bcrypt::hash(x, bcrypt::DEFAULT_COST).unwrap())),
|
||||
vk_id: Set(value.vk_id),
|
||||
telegram_id: Set(None),
|
||||
group: Set(Some(value.group)),
|
||||
role: Set(value.role),
|
||||
android_version: Set(Some(value.version)),
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -241,8 +246,6 @@ mod schema {
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use crate::database::driver;
|
||||
use crate::database::models::UserRole;
|
||||
use crate::routes::auth::sign_up::schema::Request;
|
||||
use crate::routes::auth::sign_up::sign_up;
|
||||
use crate::test_env::tests::{static_app_state, test_app_state, test_env};
|
||||
@@ -251,6 +254,10 @@ mod tests {
|
||||
use actix_web::http::Method;
|
||||
use actix_web::http::StatusCode;
|
||||
use actix_web::test;
|
||||
use database::entity::sea_orm_active_enums::UserRole;
|
||||
use database::entity::{UserColumn, UserEntity};
|
||||
use database::sea_orm::ColumnTrait;
|
||||
use database::sea_orm::{EntityTrait, QueryFilter};
|
||||
|
||||
struct SignUpPartial<'a> {
|
||||
username: &'a str,
|
||||
@@ -282,7 +289,12 @@ mod tests {
|
||||
test_env();
|
||||
|
||||
let app_state = static_app_state().await;
|
||||
driver::users::delete_by_username(&app_state, &"test::sign_up_valid".to_string()).await;
|
||||
|
||||
UserEntity::delete_many()
|
||||
.filter(UserColumn::Username.eq("test::sign_up_valid"))
|
||||
.exec(app_state.get_database())
|
||||
.await
|
||||
.expect("Failed to delete user");
|
||||
|
||||
// test
|
||||
|
||||
@@ -303,7 +315,12 @@ mod tests {
|
||||
test_env();
|
||||
|
||||
let app_state = static_app_state().await;
|
||||
driver::users::delete_by_username(&app_state, &"test::sign_up_multiple".to_string()).await;
|
||||
|
||||
UserEntity::delete_many()
|
||||
.filter(UserColumn::Username.eq("test::sign_up_multiple"))
|
||||
.exec(app_state.get_database())
|
||||
.await
|
||||
.expect("Failed to delete user");
|
||||
|
||||
let create = sign_up_client(SignUpPartial {
|
||||
username: "test::sign_up_multiple",
|
||||
|
||||
@@ -1,12 +1,13 @@
|
||||
use self::schema::*;
|
||||
use crate::database::driver;
|
||||
use crate::database::driver::users::UserSave;
|
||||
use crate::database::models::{User, UserRole};
|
||||
use crate::routes::schema::ResponseError;
|
||||
use crate::utility::telegram::{WebAppInitDataMap, WebAppUser};
|
||||
use crate::{AppState, utility};
|
||||
use crate::{utility, AppState};
|
||||
use actix_web::{post, web};
|
||||
use chrono::{DateTime, Duration, Utc};
|
||||
use database::entity::sea_orm_active_enums::UserRole;
|
||||
use database::entity::ActiveUser;
|
||||
use database::query::Query;
|
||||
use database::sea_orm::{ActiveModelTrait, Set};
|
||||
use objectid::ObjectId;
|
||||
use std::sync::Arc;
|
||||
use web::Json;
|
||||
@@ -48,39 +49,32 @@ pub async fn telegram_auth(
|
||||
let web_app_user =
|
||||
serde_json::from_str::<WebAppUser>(init_data.data_map.get("user").unwrap()).unwrap();
|
||||
|
||||
let mut user = {
|
||||
match driver::users::get_by_telegram_id(&app_state, web_app_user.id).await {
|
||||
Ok(value) => Ok(value),
|
||||
Err(_) => {
|
||||
let new_user = User {
|
||||
id: ObjectId::new().unwrap().to_string(),
|
||||
username: format!("telegram_{}", web_app_user.id), // можно оставить, а можно поменять
|
||||
password: None, // ибо нехуй
|
||||
vk_id: None,
|
||||
telegram_id: Some(web_app_user.id),
|
||||
access_token: None, // установится ниже
|
||||
group: None,
|
||||
role: UserRole::Student, // TODO: при реге проверять данные
|
||||
android_version: None,
|
||||
};
|
||||
let user = match Query::find_user_by_telegram_id(app_state.get_database(), web_app_user.id)
|
||||
.await
|
||||
.expect("Failed to find user by telegram id")
|
||||
{
|
||||
Some(value) => value,
|
||||
None => {
|
||||
let new_user = ActiveUser {
|
||||
id: Set(ObjectId::new().unwrap().to_string()),
|
||||
username: Set(format!("telegram_{}", web_app_user.id)), // можно оставить, а можно поменять
|
||||
password: Set(None), // ибо нехуй
|
||||
vk_id: Set(None),
|
||||
telegram_id: Set(Some(web_app_user.id)),
|
||||
group: Set(None),
|
||||
role: Set(UserRole::Student), // TODO: при реге проверять данные
|
||||
android_version: Set(None),
|
||||
};
|
||||
|
||||
driver::users::insert(&app_state, &new_user)
|
||||
.await
|
||||
.map(|_| new_user)
|
||||
}
|
||||
new_user
|
||||
.insert(app_state.get_database())
|
||||
.await
|
||||
.expect("Failed to insert user")
|
||||
}
|
||||
.expect("Failed to get or add user")
|
||||
};
|
||||
|
||||
user.access_token = Some(utility::jwt::encode(&user.id));
|
||||
|
||||
user.save(&app_state).await.expect("Failed to update user");
|
||||
|
||||
Ok(Response::new(
|
||||
&*user.access_token.unwrap(),
|
||||
user.group.is_some(),
|
||||
))
|
||||
.into()
|
||||
let access_token = utility::jwt::encode(&user.id);
|
||||
Ok(Response::new(&access_token, user.group.is_some())).into()
|
||||
}
|
||||
|
||||
mod schema {
|
||||
@@ -89,9 +83,9 @@ mod schema {
|
||||
use crate::utility::telegram::VerifyError;
|
||||
use actix_macros::ErrResponse;
|
||||
use actix_web::body::EitherBody;
|
||||
use actix_web::cookie::CookieBuilder;
|
||||
use actix_web::cookie::time::OffsetDateTime;
|
||||
use actix_web::{HttpRequest, HttpResponse, web};
|
||||
use actix_web::cookie::CookieBuilder;
|
||||
use actix_web::{web, HttpRequest, HttpResponse};
|
||||
use derive_more::Display;
|
||||
use serde::{Deserialize, Serialize, Serializer};
|
||||
use std::ops::Add;
|
||||
@@ -131,7 +125,7 @@ mod schema {
|
||||
&mut self,
|
||||
request: &HttpRequest,
|
||||
response: &mut HttpResponse<EitherBody<String>>,
|
||||
) -> () {
|
||||
) {
|
||||
let access_token = &self.access_token;
|
||||
|
||||
let app_state = request.app_data::<web::Data<AppState>>().unwrap();
|
||||
@@ -176,4 +170,4 @@ mod schema {
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,11 +1,11 @@
|
||||
use self::schema::*;
|
||||
use crate::AppState;
|
||||
use crate::database::driver;
|
||||
use crate::database::driver::users::UserSave;
|
||||
use crate::database::models::User;
|
||||
use crate::extractors::base::AsyncExtractor;
|
||||
use crate::routes::schema::ResponseError;
|
||||
use crate::AppState;
|
||||
use actix_web::{post, web};
|
||||
use database::entity::User;
|
||||
use database::query::Query;
|
||||
use database::sea_orm::{ActiveModelTrait, IntoActiveModel, Set};
|
||||
use web::Json;
|
||||
|
||||
#[utoipa::path(responses(
|
||||
@@ -20,7 +20,7 @@ pub async fn telegram_complete(
|
||||
app_state: web::Data<AppState>,
|
||||
user: AsyncExtractor<User>,
|
||||
) -> ServiceResponse {
|
||||
let mut user = user.into_inner();
|
||||
let user = user.into_inner();
|
||||
|
||||
// проверка на перезапись уже имеющихся данных
|
||||
if user.group.is_some() {
|
||||
@@ -29,13 +29,19 @@ pub async fn telegram_complete(
|
||||
|
||||
let data = data.into_inner();
|
||||
|
||||
let db = app_state.get_database();
|
||||
let mut active_user = user.clone().into_active_model();
|
||||
|
||||
// замена существующего имени, если оно отличается
|
||||
if user.username != data.username {
|
||||
if driver::users::contains_by_username(&app_state, &data.username).await {
|
||||
if Query::is_user_exists_by_username(db, &data.username)
|
||||
.await
|
||||
.unwrap()
|
||||
{
|
||||
return Err(ErrorCode::UsernameAlreadyExists).into();
|
||||
}
|
||||
|
||||
user.username = data.username;
|
||||
active_user.username = Set(data.username);
|
||||
}
|
||||
|
||||
// проверка на существование группы
|
||||
@@ -50,9 +56,12 @@ pub async fn telegram_complete(
|
||||
return Err(ErrorCode::InvalidGroupName).into();
|
||||
}
|
||||
|
||||
user.group = Some(data.group);
|
||||
active_user.group = Set(Some(data.group));
|
||||
|
||||
user.save(&app_state).await.expect("Failed to update user");
|
||||
active_user
|
||||
.update(db)
|
||||
.await
|
||||
.expect("Failed to update user");
|
||||
|
||||
Ok(()).into()
|
||||
}
|
||||
@@ -80,11 +89,11 @@ mod schema {
|
||||
#[serde(rename_all = "SCREAMING_SNAKE_CASE")]
|
||||
#[schema(as = Flow::TelegramFill::ErrorCode)]
|
||||
pub enum ErrorCode {
|
||||
#[display("This flow already completed.")]
|
||||
#[display("This flow is already completed.")]
|
||||
#[status_code = "actix_web::http::StatusCode::CONFLICT"]
|
||||
AlreadyCompleted,
|
||||
|
||||
#[display("Username is already exists.")]
|
||||
#[display("User with that name already exists.")]
|
||||
#[status_code = "actix_web::http::StatusCode::BAD_REQUEST"]
|
||||
UsernameAlreadyExists,
|
||||
|
||||
|
||||
@@ -1,11 +1,17 @@
|
||||
use crate::AppState;
|
||||
use crate::routes::schedule::schema::CacheStatus;
|
||||
use crate::AppState;
|
||||
use actix_web::{get, web};
|
||||
use std::ops::Deref;
|
||||
|
||||
#[utoipa::path(responses(
|
||||
(status = OK, body = CacheStatus),
|
||||
))]
|
||||
#[get("/cache-status")]
|
||||
pub async fn cache_status(app_state: web::Data<AppState>) -> CacheStatus {
|
||||
CacheStatus::from(&app_state).await.into()
|
||||
app_state
|
||||
.get_schedule_snapshot("eng_polytechnic")
|
||||
.await
|
||||
.unwrap()
|
||||
.deref()
|
||||
.into()
|
||||
}
|
||||
|
||||
@@ -1,10 +1,10 @@
|
||||
use self::schema::*;
|
||||
use crate::AppState;
|
||||
use crate::database::models::User;
|
||||
use crate::extractors::base::AsyncExtractor;
|
||||
use crate::routes::schedule::schema::ScheduleEntryResponse;
|
||||
use crate::routes::schema::ResponseError;
|
||||
use actix_web::{get, web};
|
||||
use database::entity::User;
|
||||
|
||||
#[utoipa::path(responses(
|
||||
(status = OK, body = ScheduleEntryResponse),
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
mod cache_status;
|
||||
mod group;
|
||||
mod group_names;
|
||||
mod schedule;
|
||||
mod get;
|
||||
mod schema;
|
||||
mod teacher;
|
||||
mod teacher_names;
|
||||
@@ -9,6 +9,6 @@ mod teacher_names;
|
||||
pub use cache_status::*;
|
||||
pub use group::*;
|
||||
pub use group_names::*;
|
||||
pub use schedule::*;
|
||||
pub use get::*;
|
||||
pub use teacher::*;
|
||||
pub use teacher_names::*;
|
||||
|
||||
@@ -63,18 +63,6 @@ pub struct CacheStatus {
|
||||
pub updated_at: i64,
|
||||
}
|
||||
|
||||
impl CacheStatus {
|
||||
pub async fn from(value: &web::Data<AppState>) -> Self {
|
||||
From::<&ScheduleSnapshot>::from(
|
||||
value
|
||||
.get_schedule_snapshot("eng_polytechnic")
|
||||
.await
|
||||
.unwrap()
|
||||
.deref(),
|
||||
)
|
||||
}
|
||||
}
|
||||
|
||||
impl From<&ScheduleSnapshot> for CacheStatus {
|
||||
fn from(value: &ScheduleSnapshot) -> Self {
|
||||
Self {
|
||||
|
||||
@@ -13,13 +13,13 @@ where
|
||||
E: Serialize + PartialSchema + Display + PartialErrResponse;
|
||||
|
||||
/// Transform Response<T, E> into Result<T, E>
|
||||
impl<T, E> Into<Result<T, E>> for Response<T, E>
|
||||
impl<T, E> From<Response<T, E>> for Result<T, E>
|
||||
where
|
||||
T: Serialize + PartialSchema + PartialOkResponse,
|
||||
E: Serialize + PartialSchema + Display + PartialErrResponse,
|
||||
{
|
||||
fn into(self) -> Result<T, E> {
|
||||
self.0
|
||||
fn from(value: Response<T, E>) -> Self {
|
||||
value.0
|
||||
}
|
||||
}
|
||||
|
||||
@@ -46,7 +46,7 @@ where
|
||||
{
|
||||
match &self.0 {
|
||||
Ok(ok) => serializer.serialize_some(&ok),
|
||||
Err(err) => serializer.serialize_some(&ResponseError::<E>::from(err.clone().into())),
|
||||
Err(err) => serializer.serialize_some(&err.clone().into()),
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -95,7 +95,7 @@ pub trait PartialOkResponse {
|
||||
&mut self,
|
||||
_request: &HttpRequest,
|
||||
_response: &mut HttpResponse<EitherBody<String>>,
|
||||
) -> () {
|
||||
) {
|
||||
}
|
||||
}
|
||||
|
||||
@@ -126,8 +126,9 @@ where
|
||||
}
|
||||
|
||||
pub mod user {
|
||||
use crate::database::models::{User, UserRole};
|
||||
use actix_macros::{OkResponse, ResponderJson};
|
||||
use database::entity::sea_orm_active_enums::UserRole;
|
||||
use database::entity::User;
|
||||
use serde::Serialize;
|
||||
|
||||
//noinspection SpellCheckingInspection
|
||||
@@ -165,17 +166,31 @@ pub mod user {
|
||||
pub access_token: Option<String>,
|
||||
}
|
||||
|
||||
/// Create UserResponse from User ref.
|
||||
impl From<&User> for UserResponse {
|
||||
fn from(user: &User) -> Self {
|
||||
UserResponse {
|
||||
impl UserResponse {
|
||||
pub fn from_user_with_token(user: User, access_token: String) -> Self {
|
||||
Self {
|
||||
id: user.id.clone(),
|
||||
username: user.username.clone(),
|
||||
group: user.group.clone(),
|
||||
role: user.role.clone(),
|
||||
vk_id: user.vk_id.clone(),
|
||||
telegram_id: user.telegram_id.clone(),
|
||||
access_token: user.access_token.clone(),
|
||||
vk_id: user.vk_id,
|
||||
telegram_id: user.telegram_id,
|
||||
access_token: Some(access_token),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Create UserResponse from User ref.
|
||||
impl From<&User> for UserResponse {
|
||||
fn from(user: &User) -> Self {
|
||||
Self {
|
||||
id: user.id.clone(),
|
||||
username: user.username.clone(),
|
||||
group: user.group.clone(),
|
||||
role: user.role.clone(),
|
||||
vk_id: user.vk_id,
|
||||
telegram_id: user.telegram_id,
|
||||
access_token: None,
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -183,14 +198,14 @@ pub mod user {
|
||||
/// Transform User to UserResponse.
|
||||
impl From<User> for UserResponse {
|
||||
fn from(user: User) -> Self {
|
||||
UserResponse {
|
||||
Self {
|
||||
id: user.id,
|
||||
username: user.username,
|
||||
group: user.group,
|
||||
role: user.role,
|
||||
vk_id: user.vk_id,
|
||||
telegram_id: user.telegram_id,
|
||||
access_token: user.access_token,
|
||||
access_token: None,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,9 +1,9 @@
|
||||
use self::schema::*;
|
||||
use crate::database::driver::users::UserSave;
|
||||
use crate::database::models::User;
|
||||
use crate::extractors::base::AsyncExtractor;
|
||||
use crate::state::AppState;
|
||||
use actix_web::{post, web};
|
||||
use database::entity::User;
|
||||
use database::sea_orm::{ActiveModelTrait, IntoActiveModel, Set};
|
||||
|
||||
#[utoipa::path(responses((status = OK)))]
|
||||
#[post("/change-group")]
|
||||
@@ -12,9 +12,13 @@ pub async fn change_group(
|
||||
user: AsyncExtractor<User>,
|
||||
data: web::Json<Request>,
|
||||
) -> ServiceResponse {
|
||||
let mut user = user.into_inner();
|
||||
let user = user.into_inner();
|
||||
|
||||
if user.group.is_some_and(|group| group == data.group) {
|
||||
if user
|
||||
.group
|
||||
.as_ref()
|
||||
.is_some_and(|group| group.eq(&data.group))
|
||||
{
|
||||
return Ok(()).into();
|
||||
}
|
||||
|
||||
@@ -28,10 +32,12 @@ pub async fn change_group(
|
||||
{
|
||||
return Err(ErrorCode::NotFound).into();
|
||||
}
|
||||
|
||||
user.group = Some(data.into_inner().group);
|
||||
user.save(&app_state).await.unwrap();
|
||||
|
||||
|
||||
let mut active_user = user.clone().into_active_model();
|
||||
active_user.group = Set(Some(data.into_inner().group));
|
||||
|
||||
active_user.update(app_state.get_database()).await.unwrap();
|
||||
|
||||
Ok(()).into()
|
||||
}
|
||||
|
||||
|
||||
@@ -1,10 +1,10 @@
|
||||
use self::schema::*;
|
||||
use crate::database::driver;
|
||||
use crate::database::driver::users::UserSave;
|
||||
use crate::database::models::User;
|
||||
use crate::extractors::base::AsyncExtractor;
|
||||
use crate::state::AppState;
|
||||
use actix_web::{post, web};
|
||||
use database::entity::User;
|
||||
use database::query::Query;
|
||||
use database::sea_orm::{ActiveModelTrait, IntoActiveModel, Set};
|
||||
|
||||
#[utoipa::path(responses((status = OK)))]
|
||||
#[post("/change-username")]
|
||||
@@ -13,21 +13,24 @@ pub async fn change_username(
|
||||
user: AsyncExtractor<User>,
|
||||
data: web::Json<Request>,
|
||||
) -> ServiceResponse {
|
||||
let mut user = user.into_inner();
|
||||
let user = user.into_inner();
|
||||
|
||||
if user.username == data.username {
|
||||
return Ok(()).into();
|
||||
}
|
||||
|
||||
if driver::users::get_by_username(&app_state, &data.username)
|
||||
let db = app_state.get_database();
|
||||
|
||||
if Query::is_user_exists_by_username(db, &data.username)
|
||||
.await
|
||||
.is_ok()
|
||||
.unwrap()
|
||||
{
|
||||
return Err(ErrorCode::AlreadyExists).into();
|
||||
}
|
||||
|
||||
user.username = data.into_inner().username;
|
||||
user.save(&app_state).await.unwrap();
|
||||
let mut active_user = user.into_active_model();
|
||||
active_user.username = Set(data.into_inner().username);
|
||||
active_user.update(db).await.unwrap();
|
||||
|
||||
Ok(()).into()
|
||||
}
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
use crate::database::models::User;
|
||||
use crate::extractors::base::AsyncExtractor;
|
||||
use crate::routes::schema::user::UserResponse;
|
||||
use actix_web::get;
|
||||
use database::entity::User;
|
||||
|
||||
#[utoipa::path(responses((status = OK, body = UserResponse)))]
|
||||
#[get("/me")]
|
||||
|
||||
@@ -2,29 +2,30 @@ mod env;
|
||||
|
||||
pub use crate::state::env::AppEnv;
|
||||
use actix_web::web;
|
||||
use diesel::{Connection, PgConnection};
|
||||
use database::migration::{Migrator, MigratorTrait};
|
||||
use database::sea_orm::{ConnectOptions, Database, DatabaseConnection};
|
||||
use providers::base::{ScheduleProvider, ScheduleSnapshot};
|
||||
use std::collections::HashMap;
|
||||
use std::sync::Arc;
|
||||
use tokio::sync::{Mutex, MutexGuard};
|
||||
use std::time::Duration;
|
||||
use tokio_util::sync::CancellationToken;
|
||||
|
||||
/// Common data provided to endpoints.
|
||||
pub struct AppState {
|
||||
cancel_token: CancellationToken,
|
||||
database: Mutex<PgConnection>,
|
||||
database: DatabaseConnection,
|
||||
providers: HashMap<String, Arc<dyn ScheduleProvider>>,
|
||||
env: AppEnv,
|
||||
}
|
||||
|
||||
impl AppState {
|
||||
pub async fn new() -> Result<Self, Box<dyn std::error::Error>> {
|
||||
let database_url = std::env::var("DATABASE_URL").expect("DATABASE_URL must be set");
|
||||
|
||||
pub async fn new(
|
||||
database: Option<DatabaseConnection>,
|
||||
) -> Result<Self, Box<dyn std::error::Error>> {
|
||||
let env = AppEnv::default();
|
||||
let providers: HashMap<String, Arc<dyn ScheduleProvider>> = HashMap::from([(
|
||||
"eng_polytechnic".to_string(),
|
||||
providers::EngelsPolytechnicProvider::new({
|
||||
providers::EngelsPolytechnicProvider::get({
|
||||
#[cfg(test)]
|
||||
{
|
||||
providers::EngelsPolytechnicUpdateSource::Prepared(ScheduleSnapshot {
|
||||
@@ -52,16 +53,35 @@ impl AppState {
|
||||
|
||||
let this = Self {
|
||||
cancel_token: CancellationToken::new(),
|
||||
database: Mutex::new(
|
||||
PgConnection::establish(&database_url)
|
||||
.unwrap_or_else(|_| panic!("Error connecting to {}", database_url)),
|
||||
),
|
||||
database: if let Some(database) = database {
|
||||
database
|
||||
} else {
|
||||
let database_url = std::env::var("DATABASE_URL").expect("DATABASE_URL must be set");
|
||||
|
||||
let mut opt = ConnectOptions::new(database_url.clone());
|
||||
|
||||
opt.max_connections(4)
|
||||
.min_connections(2)
|
||||
.connect_timeout(Duration::from_secs(10))
|
||||
.idle_timeout(Duration::from_secs(8))
|
||||
.sqlx_logging(true);
|
||||
|
||||
let database = Database::connect(opt)
|
||||
.await
|
||||
.unwrap_or_else(|_| panic!("Error connecting to {}", database_url));
|
||||
|
||||
Migrator::up(&database, None)
|
||||
.await
|
||||
.expect("Failed to run database migrations");
|
||||
|
||||
database
|
||||
},
|
||||
env,
|
||||
providers,
|
||||
};
|
||||
|
||||
if this.env.schedule.auto_update {
|
||||
for (_, provider) in &this.providers {
|
||||
for provider in this.providers.values() {
|
||||
let provider = provider.clone();
|
||||
let cancel_token = this.cancel_token.clone();
|
||||
|
||||
@@ -80,8 +100,8 @@ impl AppState {
|
||||
None
|
||||
}
|
||||
|
||||
pub async fn get_database(&'_ self) -> MutexGuard<'_, PgConnection> {
|
||||
self.database.lock().await
|
||||
pub fn get_database(&'_ self) -> &DatabaseConnection {
|
||||
&self.database
|
||||
}
|
||||
|
||||
pub fn get_env(&self) -> &AppEnv {
|
||||
@@ -90,6 +110,8 @@ impl AppState {
|
||||
}
|
||||
|
||||
/// Create a new object web::Data<AppState>.
|
||||
pub async fn new_app_state() -> Result<web::Data<AppState>, Box<dyn std::error::Error>> {
|
||||
Ok(web::Data::new(AppState::new().await?))
|
||||
pub async fn new_app_state(
|
||||
database: Option<DatabaseConnection>,
|
||||
) -> Result<web::Data<AppState>, Box<dyn std::error::Error>> {
|
||||
Ok(web::Data::new(AppState::new(database).await?))
|
||||
}
|
||||
|
||||
@@ -7,20 +7,31 @@ pub(crate) mod tests {
|
||||
|
||||
pub fn test_env() {
|
||||
info!("Loading test environment file...");
|
||||
|
||||
dotenvy::from_filename(".env.test.local")
|
||||
.or_else(|_| dotenvy::from_filename(".env.test"))
|
||||
.expect("Failed to load test environment file");
|
||||
}
|
||||
|
||||
pub async fn test_app_state() -> web::Data<AppState> {
|
||||
let state = new_app_state().await.unwrap();
|
||||
let state = new_app_state(Some(static_app_state().await.get_database().clone()))
|
||||
.await
|
||||
.unwrap();
|
||||
|
||||
state.clone()
|
||||
}
|
||||
|
||||
pub async fn static_app_state() -> web::Data<AppState> {
|
||||
static STATE: OnceCell<web::Data<AppState>> = OnceCell::const_new();
|
||||
|
||||
STATE.get_or_init(|| test_app_state()).await.clone()
|
||||
|
||||
STATE
|
||||
.get_or_init(async || -> web::Data<AppState> {
|
||||
#[cfg(feature = "trace")]
|
||||
console_subscriber::init();
|
||||
|
||||
new_app_state(None).await.unwrap()
|
||||
})
|
||||
.await
|
||||
.clone()
|
||||
}
|
||||
}
|
||||
|
||||
@@ -24,14 +24,13 @@ static ENCODING_KEY: LazyLock<EncodingKey> = LazyLock::new(|| {
|
||||
});
|
||||
|
||||
/// Token verification errors.
|
||||
#[allow(dead_code)]
|
||||
#[derive(Debug)]
|
||||
pub enum Error {
|
||||
/// The token has a different signature.
|
||||
InvalidSignature,
|
||||
|
||||
/// Token reading error.
|
||||
InvalidToken(ErrorKind),
|
||||
InvalidToken,
|
||||
|
||||
/// Token expired.
|
||||
Expired,
|
||||
@@ -63,13 +62,13 @@ struct Claims {
|
||||
pub(crate) const DEFAULT_ALGORITHM: Algorithm = Algorithm::HS256;
|
||||
|
||||
/// Checking the token and extracting the UUID of the user account from it.
|
||||
pub fn verify_and_decode(token: &String) -> Result<String, Error> {
|
||||
pub fn verify_and_decode(token: &str) -> Result<String, Error> {
|
||||
let mut validation = Validation::new(DEFAULT_ALGORITHM);
|
||||
|
||||
validation.required_spec_claims.remove("exp");
|
||||
validation.validate_exp = false;
|
||||
|
||||
let result = decode::<Claims>(&token, &*DECODING_KEY, &validation);
|
||||
let result = decode::<Claims>(token, &DECODING_KEY, &validation);
|
||||
|
||||
match result {
|
||||
Ok(token_data) => {
|
||||
@@ -82,13 +81,13 @@ pub fn verify_and_decode(token: &String) -> Result<String, Error> {
|
||||
Err(err) => Err(match err.into_kind() {
|
||||
ErrorKind::InvalidSignature => Error::InvalidSignature,
|
||||
ErrorKind::ExpiredSignature => Error::Expired,
|
||||
kind => Error::InvalidToken(kind),
|
||||
_ => Error::InvalidToken,
|
||||
}),
|
||||
}
|
||||
}
|
||||
|
||||
/// Creating a user token.
|
||||
pub fn encode(id: &String) -> String {
|
||||
pub fn encode(id: &str) -> String {
|
||||
let header = Header {
|
||||
typ: Some(String::from("JWT")),
|
||||
..Default::default()
|
||||
@@ -98,12 +97,12 @@ pub fn encode(id: &String) -> String {
|
||||
let exp = iat + Duration::days(365 * 4);
|
||||
|
||||
let claims = Claims {
|
||||
id: id.clone(),
|
||||
id: id.to_string(),
|
||||
iat: iat.timestamp().unsigned_abs(),
|
||||
exp: exp.timestamp().unsigned_abs(),
|
||||
};
|
||||
|
||||
jsonwebtoken::encode(&header, &claims, &*ENCODING_KEY).unwrap()
|
||||
jsonwebtoken::encode(&header, &claims, &ENCODING_KEY).unwrap()
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
@@ -115,7 +114,7 @@ mod tests {
|
||||
fn test_encode() {
|
||||
test_env();
|
||||
|
||||
assert_eq!(encode(&"test".to_string()).is_empty(), false);
|
||||
assert!(!encode("test").is_empty());
|
||||
}
|
||||
|
||||
#[test]
|
||||
@@ -128,7 +127,7 @@ mod tests {
|
||||
assert!(result.is_err());
|
||||
assert_eq!(
|
||||
result.err().unwrap(),
|
||||
Error::InvalidToken(ErrorKind::InvalidToken)
|
||||
Error::InvalidToken
|
||||
);
|
||||
}
|
||||
|
||||
|
||||
@@ -33,7 +33,7 @@ impl WebAppInitDataMap {
|
||||
};
|
||||
|
||||
data.split('&')
|
||||
.map(|kv| kv.split_once('=').unwrap_or_else(|| (kv, "")))
|
||||
.map(|kv| kv.split_once('=').unwrap_or((kv, "")))
|
||||
.for_each(|(key, value)| {
|
||||
this.data_map.insert(key.to_string(), value.to_string());
|
||||
});
|
||||
|
||||
Reference in New Issue
Block a user