NeurIPS 2023 Poster Session 4 (Thursday Morning)
Key Takeaways at a Glance
00:58
Temporal action segmentation is the task of translating an untrimmed video into action segments.02:12
Activity grammar induction algorithm and effective parser improve segmentation optimization.11:15
Combining discriminative and generative approaches improves perception.12:19
Online adaptation improves performance in streaming scenarios.15:28
The main goal is to improve performance during training.16:00
Consider the difference between online adaptation and training time adaptation.17:13
Sample complexity in recurrent neural networks.21:22
Streaming models and their applications.25:19
Matrix sketching and its role in reducing dimensionality.27:10
Making models robust to data transformations.29:41
Adapting pre-trained models using metric space information.39:28
Applying the method to different domains.42:36
Improving predictions on a subset of classes.43:06
Limitation of the argmax technique in predicting endpoints.44:51
Using distance matrix to improve prediction accuracy.46:11
Applications of feature shift detection.47:43
Iterative process for feature localization and correction.52:29
Evaluation metrics for feature correction.53:40
Selective feature replacement for better correction.55:07
Iterative process for multiple corrections.
1. Temporal action segmentation is the task of translating an untrimmed video into action segments.
🥈85
00:58
The goal is to classify the action labels of each frame and refine out-of-context errors to fit into the activity context.
- Existing models often suffer from out-of-context errors and fail to distinguish between similar actions.
- The proposed approach includes activity grammar induction and an effective parser to find optimal action sequences.
2. Activity grammar induction algorithm and effective parser improve segmentation optimization.
🥇92
02:12
The activity grammar is based on probabilistic context-free grammar and includes variables, commas, rules, and star symbols.
- The grammar helps refine out-of-context errors and can recursively generate and expand sequences.
- The proposed approach improves baseline performance and effectively removes unnecessary actions.
3. Combining discriminative and generative approaches improves perception.
🥈88
11:15
By adapting a pre-trained generator model using a diffusion loss, both discriminative ability and generalization ability can be enhanced.
- Discriminative models fit well to the training set but may lack generalization.
- Generative models better generalize but may not fit the training set perfectly.
4. Online adaptation improves performance in streaming scenarios.
🥇91
12:19
Adapting models in a streaming manner can significantly improve performance compared to traditional test-time adaptation methods.
- Online adaptation accumulates information learned from previous examples.
- Significant improvements were observed in image classification and online adaptation tasks.
5. The main goal is to improve performance during training.
🥈85
15:28
Improving performance during training is a challenging engineering problem that requires matching and improving the performance of the model.
- This approach is considered an adaptation technique due to its engineering simplicity.
- Starting with a good initialization is ideal, but training using generative and discriminative losses can achieve the best results.
6. Consider the difference between online adaptation and training time adaptation.
🥉78
16:00
Results show that classifiers can be improved through online adaptation, but there is uncertainty about the difference compared to training time adaptation.
- Results on imet and imag datasets demonstrate improved classifiers.
- However, concerns about the generator model overpowering the discriminator model and the need for regularization arise.
7. Sample complexity in recurrent neural networks.
🥈88
17:13
The work focuses on understanding the number of data points or samples needed to train a recurrent neural network.
- The number of samples needed grows linearly with the sequence length.
- Adding a small amount of noise to the network's activation values reduces the number of samples needed.
8. Streaming models and their applications.
🥈82
21:22
Streaming models are used when there is a large amount of data that cannot be stored locally.
- Examples include Netflix user ratings and online data streams.
- The streaming model allows for incremental updates to the data, reducing space complexity.
9. Matrix sketching and its role in reducing dimensionality.
🥈86
25:19
Matrix sketching techniques are used to reduce the dimensionality of problems like linear regression.
- Matrix sketching preserves pairwise distances and reduces space complexity.
- These techniques can be applied to solve common problems with lower space complexity.
10. Making models robust to data transformations.
🥇91
27:10
Large pre-trained models like ChatGPT and Segment Anything Model are not completely robust to data transformations.
- Inverted images and rotated images can cause performance drops in these models.
- A strategy to address this is to side-train a small network to canonicalize the input for the pre-trained model.
11. Adapting pre-trained models using metric space information.
🥈85
29:41
By replacing the arcmax with a weighted average in the metric space, pre-trained models can be adapted to navigate the metric space and derive predictions based on the geometry of the classes.
- This approach assumes access to metric space information about the classes.
- The weights of the softmax are used to modulate the predictions based on the geometric information.
12. Applying the method to different domains.
🥇92
39:28
The method can be applied to various domains, such as image classification and air quality prediction.
- For example, in image classification, the method can be used to improve predictions by considering the hierarchy of labels in ImageNet.
- In air quality prediction, the method can navigate the metric space to predict intermediate points based on the distances between categories.
13. Improving predictions on a subset of classes.
🥈88
42:36
The method can also be used to improve predictions on a subset of classes, even when the model is trained on a larger set of classes.
- Randomly selecting a subset of classes from ImageNet and applying the method resulted in improved accuracy compared to just predicting the arcmax.
- This approach allows for more flexibility in predicting classes beyond the original set.
14. Limitation of the argmax technique in predicting endpoints.
🥈85
43:06
The argmax technique can only predict the endpoints of a given example, which is a drawback compared to other techniques.
- This limitation restricts the accuracy of predictions.
- Other techniques, like Loki, can predict more accurately by considering intermediate classes.
15. Using distance matrix to improve prediction accuracy.
🥇92
44:51
In addition to the output distribution, the distance matrix between classes is needed to apply the technique.
- The distance matrix allows for a more precise prediction by considering the relationships between classes.
- The technique involves a simple matrix-vector product and taking the argmax of the result.
16. Applications of feature shift detection.
🥉78
46:11
Feature shift detection is useful in scenarios like sensor data analysis and data standardization in different domains.
- It can be used to detect malfunctioning sensors in a sensor network.
- It can also be used to identify and fix data standardization issues between different datasets.
17. Iterative process for feature localization and correction.
🥈88
47:43
The technique involves an iterative process of detecting and correcting features that contribute to the shift.
- The process includes training a binary classifier and evaluating the F1 score to measure the accuracy of the correction.
- The technique can be applied to various types of shifts and has shown superior performance compared to other methods.
18. Evaluation metrics for feature correction.
🥈82
52:29
The F1 score is used to evaluate the accuracy of feature correction.
- The F1 score measures the balance between true positives and false negatives in detecting and correcting features.
- Lower diversions indicate better correction performance.
19. Selective feature replacement for better correction.
🥈89
53:40
Only the features that contribute to the shift are replaced with proposals from the reference dataset.
- Features that are independent of the shift are kept unchanged.
- The technique aims to achieve a perfect correction by replacing only the corrupted features.
20. Iterative process for multiple corrections.
🥈86
55:07
The correction process can be repeated multiple times to further improve the accuracy of the correction.
- The process involves retraining the classifier and evaluating the correction performance.
- Iterations continue until the balance accuracy of the classifier reaches random chance.