Team Track_Robot's codes for 2022 DSTA TIL Competition.
- The training dataset provided will contain images of humans in various standing and fallen positions.
- You are tasked to create and train an Object Detection model to:
- Detect the bounding box coordinates of all humans in the images
- Classify the positions of all humans in the images (standing or fallen)
- You are allowed to use/include external datasets of your choice but you need to cite your sources and abide by any copyright rules stated by the data owners.
- The evaluation data will consist of images of humans in various standing and fallen positions.
-
The CV task will be evaluated using AP@[0.5:0.95] (rounded off to 4 decimal places) scores, the average AP for IoU from 0.95 to 0.95 with a step size of 0.05..
-
Your model’s outputs must be stored in a JSON file in the COCO format for evaluation. Below is a sample of an expected JSON file:
- The dataset is a collection of audio clips with expressions of the following emotions: angry, fear, happy, neutral, sad.
- Your task is to train a model to perform speech emotion recognition.
- You will be provided a training set for your use. You are allowed to include additional data to train your model.
- The evaluation data will consist of audio clips spoken in multiple languages.
- Majority of the evaluation data are in the intonation of Singapore English.
-
The NLP challenge will evaluated with F1 score (rounded off to 4 decimal places), the harmonic mean of precision and recall.
-
Ensure that the submissions are in .csv with format matching the provided sample (spreadsheet with no headers, sorted in ascending order by the first column with the correct number of entries). Below is a sample of an expected csv file: