courses:wshop:topics:tematy2024wiosna

  • Student: Przemysław Pawlik
  • Namespace in the wiki: iiif
  • The goal of the project: The goal of this project is to combine IIIF technology with a knowledge graph to provide metadata-rich descriptions via IIIF stack of technology
  • Technology: Semantic Web, Python/Java
  • Description: IIIF is a set of open APIs for getting images and their metadata from dedicated server(s) and displaying them to the user (it is of course more sophisticated, e.g., it provides rich image manipulation capabilities, but we will not dive into such details in this project). The goal of this project is to combine IIIF technology with a knowledge graph. As a result, alongside a scan of a document, the user will see a rich description about that document stored in a knowledge graph.
    • Idea for the project: (a) store images on your local filesystem, (b) use Apache Jena Fuseki for storing knowledge graph (ready-to-use standalone server with knowledge graph), (c) create some code to generate IIIF manifests with rich metadata (IIIF manifests contain every information needed to display the image and metadata), (d) use Universal Viewer to display the content (it simply gets the IIIF manifest and do all the magic), (e) evaluate how these rich metadata can be searched by the users
    • Note: The goal of this project is to show a proof-of-concept of how such a thing could work. If there will be time, we will use real data/metadata and (or this can be done in following project / Master's Thesis)
  • Links:
  • Groups and namespaces in the wiki:
    • Maciej Baczyński: yolo1
    • Aleksandra Jaroszek, Maciej Struski: yolo2
  • The goal of the project: Verify whether state-of-the-art object detection models are usable for documents (manuscripts, printed documents, music scores, etc)
  • Technology: Python, machine learning, data analysis
  • Description: The object detection models (like YOLO – probably the most popular one) are used to identify things in the images. They are usually used for photos or videos, and they achieve very good performance here. The goal of the project, however, will be to examine whether such models are capable of recognizing elements of different types of documents: manuscripts, printed documents, music scores. The expected output of the model: “This part of the scan is handwritten text. While this part of the document is a plot. And here, you have a signature”. During the project you have to: (a) identify which models can be useful for such a task, (b) find/prepare the dataset with images, (c) evaluate the models.
  • Links:
  • Groups and namespaces in the wiki:
    • Kamil Butryn: hwr1
    • Magdalena Gancarek, Klaudia Korczak: hwr2
  • The goal of the project: Explore and compare state-of-the-art handwriting recognition models/methods and benchmark datasets
  • Technology: Google Scholar/ResearchGate/reading :), Python, machine learning, data analysis
  • Description: The goal of the project is to create synthetic review of the state-of-the-art APIs/models and benchmark datasets in the area of handwriting recognition. In the second part of the project, you will create the pipeline for evaluation of such APIs/models.
  • Links:
  • Student: FIXME
  • Namespace in the wiki: FIXME
  • The goal of the project: Preparation and evaluation of the ontology
  • Technology: Semantic Web
  • Description: Work in the close collaboration with the SOLARIS team to create an ontology from the perspective of training and maintenance: model the core elements and beam lines, model their parameters, generate some instances on actual synchrotron logs.
  • Links:
  • Student: FIXME
  • Namespace in the wiki: FIXME
  • The goal of the project: Evaluation of the Loki on triplestore
  • Technology: PHP, Semantic Web
  • Description: Loki was moved from SWI-Prolog to the triplestore (see loki). As a result it should be better, faster, etc. Now, it's time to evaluate this new setup: (a) performance tests, (b) usability, (c) docs consistency.
  • Links:
  • Student: Agnieszka Felis, Mikołaj Golowski
  • Namespace in the wiki: biraffe3
  • The goal of the project: Support in BIRAFFE3 experiment
  • Technology: Python, data analysis
  • Description: The project has three phases. During the first phase, you will take part in final fixes during the pilot study (March 2024). Then, in the second part, you will help with conducting the actual experiment (scope of work to be determined). Finally, you will perform some preliminary analyses on the actual data collected in the experiment (scope of work to be determined). If the work is done with appropriate involvement, you could also become a co-author of a publication on BIRAFFE3 in Nature Scientific Data.
  • Links:
  • Group: Łukasz Chlebek, Piotr Kalita, Paweł Wąsik
  • Namespace in the wiki: emognition
  • The goal of the project: Replicate and evaluate methods and tools proposed for emotion recognition by Emognition team from PWr
  • Technology: reading :), Python, data analysis, machine learning
  • Description: Our colleagues from Emognition team deal with the same tasks as we in BIRAFFE series of experiments. In this project, we want to explore if their results are really as good as they say by following their papers/tools/methods (starting point is indicated by the Links below) and trying to replicate their results. In the subsequent projects (or Master Thesis) there will be possibility to use these methods and tools to other datasets like BIRAFFE2/BIRAFFE3.
  • Links:
  • Student: FIXME
  • Namespace in the wiki: FIXME
  • The goal of the project: Evaluate usefulness of the RegFlow method for 2-D emotion prediction task
  • Technology: Python, data analysis
  • Description: Continuation of the regflow project (where RegFlow was successfully run and adapted to 2-D emotion space). The goal of the project is: (a) to train emotion prediction model(s) on the subset of the BIRAFFE2 and use RegFlow to show emotion predictions generated by the model(s), (b) to show the blobs as some kind of blobs (not as a cloud of points), © to evaluate the usefulness of such a solution.
  • Links:
  • Student: Jan Malczewski
  • Namespace in the wiki: emolatent
  • The goal of the project: Create a prediction model that maps features into artificial X-dimensional space. Then, map this space into actual emotions/labels space
  • Technology: Python, machine learning, data analysis
  • Description: Emotions in computational models are often represented as 2-dimensional Valence x Arousal space (sometimes it is a 3-D space with Dominance). When it comes to train prediction models, one try to map features space (features of physiological signals) into this 2-D space. A lot of research shows that this does not work. So, we want to try another approach: create a classifier that maps features into X-dimensional artificial space (with X » 2), and then create some mapping from this space into the Valence x Arousal space. As X will be higher than 2 it should catch more sophisticated relations in the data. A mapping from X-dimensional space into 2-D space should also provide some information about accuracy of the prediction/mapping.
    You can start with the DEAP dataset, but ultimately, the project should be done with BIRAFFE2 dataset.
    There should be some “classical” model trained/downloaded as a baseline (to show that the new model is better).
    Transformers should be useful to generate the artificial X-D space.
  • Links:
  • Student: Klaudia Ropel
  • Namespace in the wiki: emospeech
  • The goal of the project: Review of the state-of-the-art + evaluation of the solutions found on the selected subset of benchmark datasets
  • Technology: Google Scholar/ResearchGate/reading ;), Python, machine learning, data analysis
  • Description: The goal of the project is to create synthetic review of the state-of-the-art APIs/models and benchmark datasets in the area of emotion recognition from speech (both signal characteristic-based and spoken text-based). In the second part of the project, you will create the pipeline for evaluation of the speech-based emotion recognition APIs/models, i.e. prepare the dataset, predict the emotion based on sample, evaluate the prediction correctness, compare the accuracy of APIs/models under evaluation.
  • Links:
    • Starting point: Speech emotion recognition: Emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers DOI:10.1016/j.specom.2019.12.001
  • Student: FIXME
  • Namespace in the wiki: FIXME
  • The goal of the project: Review of existing datasets and verification whether Dominance is really (in)dependent factor
  • Technology: Google Scholar, Python, data analysis
  • Description: FIXME
  • Links:
  • Student: FIXME
  • Namespace in the wiki: cfeval
  • The goal of the project: The goal of this project is to implement a Python module that will cover all of the evalaution metrics from: Counterfactual explanations and how to find them: literature review and benchmarking
  • Technology: Python
  • Description: Counterfactual explanations are hypothetical instances that explain how changing certain features of an input would lead to a different model prediction. For example, in the context of image classification, a counterfactual explanation might demonstrate how modifying the color of an object in an image would cause a model to classify it as a different category, or in a credit scoring system, a counterfactual explanation could demonstrate how increasing a borrower's income and decreasing their debt-to-income ratio would result in an approval instead of a rejection for a loan application. There are multiple different methods for constructing counterfactual, and many methods to evaluate them, but no unified framework for doing that.
  • Links:
  • Student: Monika Rasz
  • Namespace in the wiki: tsproto1
  • The goal of the project: Reproduce results from a selected scientific paper describing prototypical model for time-series.
  • Technology: Python, PyTorch, Keras
  • Description: Prototypes in explainable AI are representative instances that encapsulate the typical characteristics of a given class or concept. They serve as exemplars to explain model predictions by illustrating the features most salient for classification or decision-making.In a healthcare scenario, prototypes could represent typical patient profiles within specific disease categories. For instance, a prototype for heart disease might exhibit high blood pressure, elevated cholesterol levels, and a family history of cardiovascular issues, serving as a representative example to explain why a patient's symptoms align with a heart disease diagnosis. The goal is to implement (reproduce) results from current SOTA methods for prototypical neural networks/models in time-series domain.
  • Links:
  • Student: Sebastian Sęczyk,Jakub Pleśniak
  • Namespace in the wiki: Sebastian Sęczyk, Jakub Pleśniak
  • The goal of the project: Tool for data labeling supported with eyetracking
  • Technology: python,tobii
  • Description: Eye tracking is the process of measuring and recording the movement of a person's gaze or eye movements to understand visual attention and cognitive processes. The aim is to develop a data labeling system integrating eye tracking, augmenting labels with inferred “reasons” by tracking user gaze to text, images, or data fragments associated with the label.
  • Links:
  • Student: FIXME
  • Namespace in the wiki: openmlds
  • The goal of the project: Prepare a script that will build meta-learnign dataset out of OpenML logs
  • Technology: Python, OpenML API
  • Description: The main goal of the project is to create a script that will fetch all of the runs/pipelines and dataset from OpenML platform and create a dataset out of it. The challenge is to transform pipeline definitions which are code snippets into logical components of machine-learning pipeline (including deep neural networks). Such a dataset will serve as a learn-to-learn dataset for meta-learning solutions.

  • Student: Kamil Piskorz
  • Namespace in the wiki: windowshap
  • The goal of the project: The goal is to improve the SHAP XAI algorithm to work more efficiently on time-seris data
  • Technology: Python, Keras/PyTorch
  • Description: The objective is to thoroughly test and evaluate, with potential for improvement, the implementation of window-shap for explaining time series data.
  • Links:
  • Student: Jakub Samel
  • Namespace in the wiki: start
  • The goal of the project:
  • Technology: Python, Keras/PyTorch
  • Description: Red-teaming of ML models involves assessing the robustness and vulnerabilities of a machine learning system by intentionally challenging it with adversarial attacks or realistic/counterfactual scenarios to identify weaknesses and improve its resilience to potential threats. This process helps enhance the security and reliability of the model by uncovering and addressing potential weaknesses in its design or implementation.
  • Links:
  • Student: FIXME
  • Namespace in the wiki: FIXME
  • The goal of the project: The goal is to improve LUX software to perform calculation in reduced dimensionality space
  • Technology: Python, Keras/PyTorch
  • Description: LUX (Local Universal Rule-Based Explainer) is an XAI algorithm that produces explanations for any type of machine-learning model. It provides local explanations in a form of human-readable (and executable) rules, but also provide counterfactual explanations as well as visualization of the explanations. It creates explanations by selection of neighborhood data-points which is computationally intensive as it is based on clustering algorithms. In high dims spaces this is inefficient and has limited practical usage due to dimensionality curse. The goal would be to add dimensionality reduction step to the process and test efficiency improvements.
  • Links:
  • Student: FIXME
  • Namespace in the wiki: FIXME
  • The goal of the project: FIXME
  • Technology: FIXME
  • Description: FIXME
  • Links:
    • FIXME
  • courses/wshop/topics/tematy2024wiosna.txt
  • Last modified: 8 months ago
  • by admin