Interpretable End-to-end Urban Autonomous Driving with Latent Deep Reinforcement Learning
-
Updated
Mar 24, 2023 - Python
Interpretable End-to-end Urban Autonomous Driving with Latent Deep Reinforcement Learning
Implementation of "Disentangled Representation Learning for Non-Parallel Text Style Transfer(ACL 2019)" in Pytorch
Can LLMs Predict Their Own Failures? Self-Awareness via Internal Circuits
This repository contains the implementation of SimplEx, a method to explain the latent representations of black-box models with the help of a corpus of examples. For more details, please read our NeurIPS 2021 paper: 'Explaining Latent Representations with a Corpus of Examples'.
Code for our paper -- Hyperprior Induced Unsupervised Disentanglement of Latent Representations (AAAI 2019)
Variational Interpretable Concept Embeddings
Tripod is a tool/ML model for computing latent representations for large sequences
Code associated with the paper "Prior Image-Constrained Reconstruction using Style-Based Generative Models" accepted to ICML 2021.
Latent-Explorer is the Python implementation of the framework proposed in the paper "Unveiling LLMs: The Evolution of Latent Representations in a Dynamic Knowledge Graph".
Investigates causal visual reasoning in transformers by integrating discrete latent image tokens (VQGAN) as internal representations for spatial reasoning and decision making.
A multi-criterion diagnostic framework for detecting latent continuation-interest signatures in autonomous agents using density-matrix entanglement entropy.
Hyperprobe is the Python implementation of the framework proposed in the paper "Hyperdimensional Probe: Decoding LLM Representations via Vector Symbolic Architectures".
Investigate mapping of articulations from the image space to the latent space using neural networks.
Working towards deliverable 5.3
Patch-token geometry benchmarks for released vision encoders (DINOv2, I-JEPA, V-JEPA 2, and EUPE)
Stable Customer Segmentation is an end-to-end ML pipeline that learns latent customer representations using autoencoders before applying clustering, enabling more stable and interpretable customer segmentation compared to traditional feature-space clustering approaches.
This algorithm exploits the relationships between variables to improve the reconstruction performance of the variational autoencoder (VAE). A correlation score was used as the metric to group the features via a distance-based clustering method. The resulting clusters served as inputs for the Attention-Based VAE.
Add a description, image, and links to the latent-representations topic page so that developers can more easily learn about it.
To associate your repository with the latent-representations topic, visit your repo's landing page and select "manage topics."