Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Kubernetes #103

Merged
merged 136 commits into from
Jan 19, 2022
Merged

Kubernetes #103

merged 136 commits into from
Jan 19, 2022

Conversation

mazhurin
Copy link
Collaborator

  • Kubernetes deployment
  • Incident detector
  • Saving incidents request sets into cloud storage using s3 interface
  • Streaming challenges directly to ElasticSearch

mazhurin and others added 30 commits June 4, 2020 18:34
Unit tests and linting fixes.
… configuration istead of full python model names.
…e.training.model refers to the full module path of the Model class(not the Enum key).
Country and Host features. Stratified sampling(parameter 'max_samples_per_host'). Support for nested features in JSON parser(geoip feature fix).
mkaranasou and others added 29 commits April 17, 2021 14:32
* IPCache
* Two ip caches: passed and pending
* Docker image with unit tests complete.
* The new version of spark-iforest
* Kafka is up and running
* Spark secret. 
* spark encryption configuration added
* spark ssl for ui, standalone and history config.
* self.__cache.count() replaced with head()
* send to kafka by partition id
* bug fix in accummulator reset
* disable broadcast spark.sql.autoBroadcastJoinThreshold
* Sliding window in postprocesing is optional. Set sliding_window config param to zero to disable sliding_window(defaut is also zero now)
* Redis merge logging count() before and after
* Postprocessing: rollback to sending challenge with collect() due to performance issues
* append mode for Redis write.
* Fix for default 'challenged' column set to 0.
* Spark worker Dockerfile
* pyspark bump to 2.4.7
* Argo dockerfile moved to dockerfiles folder.
* spark 2.4.6
* S3 support. Alternating cache files instead of renaming.
* No s3 deletion.
* use_storage option for request_set_cache
* Support for using kafka for sensitive data.
* Support for send_by_partitions for sensitive data.
* Support for github raw configs. Support for storing sensitive data in Kafka.
* Git config support with ssh.
* Kubernetes client mode spark deployment.
* White list urls from dashboard url link.
* Whitelisting IPs in preprocessing
* Whitelisting IPs without UDF

Co-authored-by: Maria Karanasou <karanasou@gmail.com>
client_mode in the config.
Separate clearing_house connection.
Fixing whitelist ips. Switched to left_anti join.
* Kubernetes deployment. Not finished.
* Jupyter notebook with spark.
* Kafka ACL
* DB reader
* SQL based incident detector
* Attack detection and chunks removed from AttackDetection task.
* Incident detector added.
* Incident Labeler class. Tested in Jupyter notebook.
* Optional scaling in AnomalyModel
* Some fixes in Jupyter notebooks.
start in whitelist urls
fix in sending to kafka
@mazhurin mazhurin merged commit e48390c into master Jan 19, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants