Jakob Bach
Jakob-Bach
Data Scientist | Former researcher at KIT, Karlsruhe.
Languages
Loading contributions...
Top Repositories
An R package containing dissimilarity measures and aggregation functions for time series, plus cluster validity indices. Part of the code for the paper "Understanding the Effects of Temporal Energy-Data Aggregation on Clustering Quality".
Code for the papers "Finding Optimal Diverse Feature Sets with Alternative Feature Selection" and "Alternative Feature Selection with User Control".
Code and text for the paper "A Comprehensive Study of k-Portfolios of Recent SAT Solvers".
Code for the papers "Using Constraints to Discover Sparse and Alternative Subgroup Descriptions" and "Subgroup Discovery with Small and Alternative Feature Sets".
Demonstration of some MIP/SAT/SMT solvers/optimizers in multiple programming languages.
Text for the dissertation "Leveraging Constraints for User-Centric Feature Selection".
Repositories
26Code for the papers "Finding Optimal Diverse Feature Sets with Alternative Feature Selection" and "Alternative Feature Selection with User Control".
Code for the papers "Using Constraints to Discover Sparse and Alternative Subgroup Descriptions" and "Subgroup Discovery with Small and Alternative Feature Sets".
Text for the dissertation "Leveraging Constraints for User-Centric Feature Selection".
Text for the papers "Using Constraints to Discover Sparse and Alternative Subgroup Descriptions" and "Subgroup Discovery with Small and Alternative Feature Sets".
Text for the papers "Finding Optimal Diverse Feature Sets with Alternative Feature Selection" and "Alternative Feature Selection with User Control".
Demonstration of some MIP/SAT/SMT solvers/optimizers in multiple programming languages.
An R package containing dissimilarity measures and aggregation functions for time series, plus cluster validity indices. Part of the code for the paper "Understanding the Effects of Temporal Energy-Data Aggregation on Clustering Quality".
Code for the paper "An Empirical Evaluation of Constrained Feature Selection".
Defense presentation for the dissertation "Leveraging Constraints for User-Centric Feature Selection".
Code for the paper "Analyzing and Predicting Verification of Data-Aware Process Models -- a Case Study with Spectrum Auctions".
Code for Section 2 of the paper "Data-driven exploration and continuum modeling of dislocation networks".
Code for the paper "Understanding the Effects of Temporal Energy-Data Aggregation on Clustering Quality".
Code and text for the paper "A Comprehensive Study of k-Portfolios of Recent SAT Solvers".
The supervisor repo for the "Data Science Laboratory Course" at the Karlsruhe Institute of Technology (KIT), summer term 2023.
Code and text for the suspended paper/project "Predicting Satisfiability of Benchmark Instances".
Code for the suspended paper/project "Meta-Learning Feature Importance".
Code for the suspended research project "Tournament Feature Selection".
The supervisor repo for task 2 of the "Analyzing Big Data Laboratory Course" at the Karlsruhe Institute of Technology (KIT), summer term 2019.
The supervisor repo for task 1 of the "Analyzing Big Data Laboratory Course" at the Karlsruhe Institute of Technology (KIT), summer term 2019.
The supervisor repo for the "Analyzing Big Data Laboratory Course" at the Karlsruhe Institute of Technology (KIT), summer term 2020.
The programming exercises for the lecture "Big Data Analytics" at the Karlsruhe Institute of Technology (KIT), winter term 2020/2021.
The supervisor repo for the "Data Science Laboratory Course" at the Karlsruhe Institute of Technology (KIT), summer term 2021.
The exercises for the lecture "Data Science I" at the Karlsruhe Institute of Technology (KIT), winter term 2021/2022.
The supervisor repo for the "Data Science Laboratory Course" at the Karlsruhe Institute of Technology (KIT), summer term 2022.
A group repo for the practical course "Model-Driven Software Development" at the Karlsruhe Institute of Technology (KIT), summer term 2016.
Besides its original purpose, demonstrates how to create a REST API with R and run it as a (parallelized) service.