SlideShare ist ein Scribd-Unternehmen logo
1 von 35
Downloaden Sie, um offline zu lesen
Pedestrian Behavior/Intention
Modeling for Autonomous Driving VI
Yu Huang
Yu.huang07@gmail.com
Sunnyvale, California
Outline
• CoMoGCN: Coherent Motion Aware Trajectory Prediction with Graph Representation (5.5)
• STINet: Spatio-Temporal-Interactive Network for Pedestrian Detect. and Trajectory Pred.
• AC-VRNN: Attentive Conditional-VRNN for Multi-Future Trajectory Prediction (5.17)
• Spatio-Temporal Graph Transformer Networks for Pedestrian Trajectory Prediction (5.18)
• Intention-aware Residual Bidirectional LSTM for Long-term Pedestrian Trajectory
Prediction (6.30)
• It Is Not the Journey but the Destination: Endpoint Conditioned Trajectory Prediction (7.6)
• Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian Trajectory Prediction by
Learning Adaptive Neighborhoods (7.8)
• Probabilistic Crowd GAN: Multimodal Pedestrian Trajectory Prediction using a Graph
Vehicle-Pedestrian Attention Network (7.12)
CoMoGCN: Coherent Motion Aware Trajectory
Prediction with Graph Representation
• Forecasting human trajectories is critical for tasks such as robot crowd navigation and
autonomous driving.
• Modeling social interactions is of great importance for accurate group-wise motion
prediction.
• However, most existing methods do not consider information about coherence within
the crowd, but rather only pairwise interactions.
• A framework, coherent motion aware graph convolutional network (CoMoGCN), for
trajectory prediction in crowded scenes with group constraints.
• First, cluster pedestrian trajectories into groups according to motion coherence.
• Then, use graph convolutional networks to aggregate crowd information efficiently.
• The CoMoGCN also takes advantage of variational autoencoders to capture the
multimodal nature of the human trajectories by modeling the distribution.
CoMoGCN: Coherent Motion Aware Trajectory
Prediction with Graph Representation
System overview. procedures: 1. obtain coherent motion labels for each human in an offline data pre-processing
procedure. 2. Based on the coherent motion labels for each human, establish graphs capturing intergroup and
intragroup relationships. The encoder LSTM takes past trajectories as input and feeds the encoded features into two
GCNs. 3. The embeddings from the two GCNs are concatenated and forwarded to an MLP to create a distribution.
Then, features are sampled from the distribution and fed into a decoder LSTM for trajectory prediction.
CoMoGCN: Coherent Motion Aware Trajectory
Prediction with Graph Representation
STINet: Spatio-Temporal-Interactive Network for
Pedestrian Detection and Trajectory Prediction
• Detecting pedestrians and predicting future trajectories for them are critical tasks for
numerous applications, such as autonomous driving.
• Previous methods either treat the detection and prediction as separate tasks or simply
add a trajectory regression head on top of a detector.
• An end-to-end two-stage network: Spatio-Temporal-Interactive Network (STINet).
• In addition to 3D geometry modeling of pedestrians, model the temporal information for
each of the pedestrians.
• It predicts both current and past locations in the first stage, so that each pedestrian can
be linked across frames and the comprehensive spatio-temporal information can be
captured in the second stage.
• Also, model the interaction among objects with an interaction graph, to gather the
information among the neighboring objects.
• Comprehensive experiments on the Lyft Dataset and the recently released large-scale
Waymo Open Dataset for both object detection and future trajectory prediction.
STINet: Spatio-Temporal-Interactive Network for
Pedestrian Detection and Trajectory Prediction
The overview. It takes a sequence of point clouds as input, detects pedestrians and predicts their future
trajectories simultaneously. The point clouds are processed by Pillar Feature Encoding to generate Pillar
Features. Then each Pillar Feature is fed into a backbone ResUNet to get backbone features. A Temporal
Region Proposal Network (T-RPN) takes backbone features and generated temporal proposal with past
and current boxes for each object. Spatio-Temporal-Interactive (STI) Feature Extractor learns features
for each temporal proposal which are used for final detection and trajectory prediction.
STINet: Spatio-Temporal-Interactive Network for
Pedestrian Detection and Trajectory Prediction
Backbone. Upper: overview of the backbone. The
input point cloud sequence is fed to Voxelization and
Point net to generate pseudo images, which are then
processed by ResNet U-Net to generate final
backbone feature sequence. Lower: detailed design
of ResNet U-Net.
STINet: Spatio-Temporal-Interactive Network for
Pedestrian Detection and Trajectory Prediction
Spatial-Temporal-Interactive Feature Extractor
(STI- FE): Local geometry, local dynamic and
history path features are extracted given a
temporal proposal. For local geometry and
local dynamics features, the yellow areas are
used for feature extraction. Relational
reasoning is performed across proposals’ local
features to generate interactive features.
STINet: Spatio-Temporal-Interactive Network for
Pedestrian Detection and Trajectory Prediction
AC-VRNN: Attentive Conditional-VRNN for
Multi-Future Trajectory Prediction
• Anticipating human motion in crowded scenarios is essential for developing intelligent
transportation systems, social-aware robots and advanced video-surveillance
applications.
• An important aspect of such task is represented by the inherently multi-modal nature of
human paths which makes socially-acceptable multiple futures when human interactions
are involved.
• A generative model for multi-future trajectory prediction based on Conditional
Variational Recurrent Neural Networks (C-VRNNs).
• Conditioning relies on prior belief maps, representing most likely moving directions and
forcing the model to consider the collective agents’ motion.
• Human interactions are modeled in a structured way with a graph attention mechanism,
providing an online attentive hidden state refinement of the recurrent estimation.
• Compared to sequence-to-sequence methods, this model operates step- by-step,
generating more refined and accurate predictions.
AC-VRNN: Attentive Conditional-VRNN for
Multi-Future Trajectory Prediction
trajectory prediction framework for a single time-step. The overall model is composed of a training module (left)
and an inference module (right). The former is composed of a recurrent variational autoencoder conditioned on
prior belief maps. The hidden state of the RNN is refined with an attentive module for the next step of
recurrence. The latter performs the displacements generation through the prior network on hidden states and
makes an online computation of the adjacency matrix which defines connections between pairs of nodes.
AC-VRNN: Attentive Conditional-VRNN for
Multi-Future Trajectory Prediction
Scheme of the attentive hidden state refinement process. The adjacency matrix is an irregular block matrix where each
block size is defined by the number of pedestrians in the current scene (a). Belief map during training for one sample
using heat similarity-based strategy. The map is centred at t − 1 to display the sampled displacements distribution at t (b).
AC-VRNN: Attentive Conditional-VRNN for
Multi-Future Trajectory Prediction
Spatio-Temporal Graph Transformer Networks for
Pedestrian Trajectory Prediction
• Understanding crowd motion dynamics is critical to real- world applications, e.g.,
surveillance systems and autonomous driving.
• This is challenging because it requires effectively modeling the socially aware crowd
spatial interaction and complex temporal dependencies.
• attention is the most important factor for trajectory prediction.
• STAR, a Spatio-Temporal grAph tRans- former framework, tackles trajectory prediction
by only attention mechanisms. STAR models intra-graph crowd interaction by TGConv, a
Transformer-based graph convolution mechanism.
• The inter-graph temporal dependencies modeled by separate temporal Transformers.
• STAR captures complex spatio-temporal interactions by interleaving between spatial and
temporal Transformers.
• To calibrate the temporal prediction for the long-lasting effect of disappeared
pedestrians, apply a read-writable external memory module, consistently being updated
by the temporal Transformer.
Spatio-Temporal Graph Transformer Networks for
Pedestrian Trajectory Prediction
STAR models the crowd as a graph and learns spatio-temporal interaction of the crowd motion
by interleaving between a graph-based spatial Transformer and a temporal Transformer
Spatio-Temporal Graph Transformer Networks for
Pedestrian Trajectory Prediction
Temporal Transformer and Spatial Transformer. (a) Temporal Transformer treats each
pedestrians independently and extracts the temporal dependencies by Transformer model (h
is the embedding of pedestrian positions, Q, K and V are the query, key, value matrix in
Transformers). (b) Spatial Transformer models the crowd as a graph, and applies TGConv, a
Transformer-based message passing graph convolution, to model the social interactions (mi→j
is the message from node i to j represented by Transformer attention)
Spatio-Temporal Graph Transformer Networks for
Pedestrian Trajectory Prediction
Network structure of STAR with application to trajectory prediction. In STAR, trajectory prediction is
achieved completely by attention mechanisms. STAR inter- leaves spatial Transformer and temporal
Transformer in two encoder blocks to extract spatio-temporal pedestrian dependencies. An external
read-writable graph memory module helps to smooth the graph embeddings and improve the
consistency of temporal predictions. The prediction at Tobs + 1 is added back to history to predict the
pedestrian poses at Tobs + 2.
Spatio-Temporal Graph Transformer Networks for
Pedestrian Trajectory Prediction
Intention-aware Residual Bidirectional LSTM for
Long-term Pedestrian Trajectory Prediction
• Trajectory prediction is one of the key capabilities for robots to safely navigate and interact with
pedestrians.
• Critical insights from human intention and behavioral patterns need to be effectively integrated
into long-term pedestrian behavior forecasting.
• An intention-aware motion prediction framework consists of a Residual Bidirectional LSTM (ReBiL)
and a mutable intention filter.
• Instead of learning step-wise displacement, learning offset to warp a nominal intention-aware
linear prediction, giving residual learning a physical intuition.
• The intention filter is inspired by genetic algorithms and particle filtering, where particles mutate
intention hypotheses throughout the pedestrian’s motion with ReBiL as the motion model.
• Experiments on a publicly available dataset under abnormal intention-changing scenarios.
Intention-aware Residual Bidirectional LSTM for
Long-term Pedestrian Trajectory Prediction
Overview of our motion prediction framework. ReBiL (dashed-line arrow) performs both truncated
prediction for particle weight update and long-term prediction at t after mutation. Mutable intention filter
takes truncated prediction results to update particle weights, and it implements Sequential Importance
Resampling (SIR) and mutation mechanism.
Intention-aware Residual Bidirectional LSTM for
Long-term Pedestrian Trajectory Prediction
It Is Not the Journey but the Destination: Endpoint
Conditioned Trajectory Prediction
• Human trajectory forecasting with multiple socially interacting agents is of
critical importance for autonomous navigation in human environments,
e.g., for self-driving cars and social robots.
• Predicted Endpoint Conditioned Network (PECNet) for flexible human
trajectory prediction.
• PECNet infers distant trajectory endpoints to assist in long-range multi-
modal trajectory prediction.
• A non- local social pooling layer enables PECNet to infer diverse yet socially
compliant trajectories.
• Additionally, a simple “truncation- trick” for improving few-shot multi-
modal trajectory prediction performance.
• Code https://karttikeya.github.io/publication/htf/
It Is Not the Journey but the Destination: Endpoint
Conditioned Trajectory Prediction
Architecture of PECNet: PECNet uses past
history, along with ground truth endpoint
to train a VAE for multi-modal endpoint
inference. Ground-truth endpoints are
denoted by ⋆ whereas x denote the
sampled endpoints Gc. The sampled
endpoints condition the social-pooling &
predictor networks for multi-agent multi-
modal trajectory forecasting. Red
connections denote the parts utilized only
during training. Shades of the same color
denote spatio-temporal neighbours
encoded with the block diagonal social
mask in social pooling module
It Is Not the Journey but the Destination: Endpoint
Conditioned Trajectory Prediction
Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian
Trajectory Prediction by Learning Adaptive Neighborhoods
• Pedestrian trajectory prediction is a prominent research track that has advanced towards
modelling of crowd social and contextual interactions, with extensive usage of Long
Short-Term Memory (LSTM) for temporal representation of walking trajectories.
• Existing approaches use virtual neighborhoods as a fixed grid for pooling social states of
pedestrians with tuning process that controls how social interactions are being captured.
• This entails performance customization to specific scenes but lowers the generalization
capability of the approaches.
• Grid-LSTM, a recent extension of LSTM, which operates over multidimensional feature
inputs.
• A perspective to interaction modeling by proposing that pedestrian neighborhoods can
become adaptive in design.
• Grid-LSTM as an encoder to learn about potential future neighborhoods and their
influence on pedestrian motion given the visual and the spatial boundaries.
• The experiment results clearly illustrate the generalization of our approach across
datasets.
Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian
Trajectory Prediction by Learning Adaptive Neighborhoods
The static neighborhood grid fO segments the
scene image into several local regions. The
dynamic grid fS takes pedestrians trajectories x1,x2
along with their looking angle to stem their
social interactions. The output static grid has
few highlighted areas, which indicates future
neighborhoods where pedestrians would walk.
Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian
Trajectory Prediction by Learning Adaptive Neighborhoods
Full pipeline of G2K kernel. The SRI
network encodes Vislets and positional
trajectories for each pedestrian
trajectory. Then maps them into social
grid mask using NLSTMv. The GNN
network discretize static context using
NLSTMo into ’Visuospatial’
neighborhoods and stores pedestrian
contextual awareness in fO. At the
consequent step, SRI takes fO and fS,
and maps them into the weighted
adjacency matrix. This will generate
the edge set ν as means of completing
graph at time-step t.
Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian
Trajectory Prediction by Learning Adaptive Neighborhoods
Gated Neighborhood Network pipeline. At the beginning, 2DCONV encodes a static image of the
scene and forward the features into NLSTM cell which discretizes the environment into a virtual grid.
Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian
Trajectory Prediction by Learning Adaptive Neighborhoods
Probabilistic Crowd GAN: Multimodal Pedestrian Trajectory
Prediction using a Graph Vehicle-Pedestrian Attention Network
• Understanding and predicting the intention of pedestrians is essential to enable
autonomous vehicles and mobile robots to navigate crowds.
• This problem becomes increasingly complex when we consider the uncertainty and
multimodality of pedestrian motion, as well as the implicit interactions between
members of a crowd, including any response to a vehicle.
• Probabilistic Crowd GAN, extends recent work in trajectory prediction, combining
Recurrent Neural Networks (RNNs) with Mixture Density Networks (MDNs) to output
probabilistic multimodal predictions, from which likely modal paths are found and used
for adversarial training.
• use of Graph Vehicle-Pedestrian Attention Network (GVAT), which models social
interactions and allows input of a shared vehicle feature, showing that inclusion of this
module leads to improved trajectory prediction both with and without the presence of a
vehicle.
• Through evaluation on various datasets, illustrates how the true multimodal and
uncertain nature of crowd interactions can be directly modelled.
Probabilistic Crowd GAN: Multimodal Pedestrian Trajectory
Prediction using a Graph Vehicle-Pedestrian Attention Network
Observed pedestrian trajectories are passed to the Generator’s encoder LSTM, whilst the relative
position of all agents, including any vehicle, are passed to the GVAT Pooling module. The
Generator outputs a GMM for each agent, from which the MultiPAC module finds the likely
modal paths, which are compared to ground truth paths by the Discriminator.
Probabilistic Crowd GAN: Multimodal Pedestrian Trajectory
Prediction using a Graph Vehicle-Pedestrian Attention Network
Node features of agent i (red) in GVAT. The
distance from i to the vehicle is appended to
each other ped-ped distance input before
encoding to account for the impact of the
vehicle on i’s relationships within the graph
Probabilistic Crowd GAN: Multimodal Pedestrian Trajectory
Prediction using a Graph Vehicle-Pedestrian Attention Network
Pedestrian Behavior/Intention Modeling for Autonomous Driving VI

Weitere ähnliche Inhalte

Was ist angesagt?

Driving behaviors for adas and autonomous driving XII
Driving behaviors for adas and autonomous driving XIIDriving behaviors for adas and autonomous driving XII
Driving behaviors for adas and autonomous driving XIIYu Huang
 
Prediction and planning for self driving at waymo
Prediction and planning for self driving at waymoPrediction and planning for self driving at waymo
Prediction and planning for self driving at waymoYu Huang
 
Deep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data IIDeep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data IIYu Huang
 
Pedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving VPedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving VYu Huang
 
Driving Behavior for ADAS and Autonomous Driving VIII
Driving Behavior for ADAS and Autonomous Driving VIIIDriving Behavior for ADAS and Autonomous Driving VIII
Driving Behavior for ADAS and Autonomous Driving VIIIYu Huang
 
Driving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VIIDriving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VIIYu Huang
 
Deep VO and SLAM
Deep VO and SLAMDeep VO and SLAM
Deep VO and SLAMYu Huang
 
Driving Behavior for ADAS and Autonomous Driving III
Driving Behavior for ADAS and Autonomous Driving IIIDriving Behavior for ADAS and Autonomous Driving III
Driving Behavior for ADAS and Autonomous Driving IIIYu Huang
 
Lidar for Autonomous Driving II (via Deep Learning)
Lidar for Autonomous Driving II (via Deep Learning)Lidar for Autonomous Driving II (via Deep Learning)
Lidar for Autonomous Driving II (via Deep Learning)Yu Huang
 
BEV Semantic Segmentation
BEV Semantic SegmentationBEV Semantic Segmentation
BEV Semantic SegmentationYu Huang
 
Camera-Based Road Lane Detection by Deep Learning II
Camera-Based Road Lane Detection by Deep Learning IICamera-Based Road Lane Detection by Deep Learning II
Camera-Based Road Lane Detection by Deep Learning IIYu Huang
 
Depth Fusion from RGB and Depth Sensors IV
Depth Fusion from RGB and Depth Sensors  IVDepth Fusion from RGB and Depth Sensors  IV
Depth Fusion from RGB and Depth Sensors IVYu Huang
 
Fisheye-Omnidirectional View in Autonomous Driving III
Fisheye-Omnidirectional View in Autonomous Driving IIIFisheye-Omnidirectional View in Autonomous Driving III
Fisheye-Omnidirectional View in Autonomous Driving IIIYu Huang
 
Deep Learning’s Application in Radar Signal Data
Deep Learning’s Application in Radar Signal DataDeep Learning’s Application in Radar Signal Data
Deep Learning’s Application in Radar Signal DataYu Huang
 
Simulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atgSimulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atgYu Huang
 
Jointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planningJointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planningYu Huang
 
3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving IIYu Huang
 
camera-based Lane detection by deep learning
camera-based Lane detection by deep learningcamera-based Lane detection by deep learning
camera-based Lane detection by deep learningYu Huang
 
Deep VO and SLAM IV
Deep VO and SLAM IVDeep VO and SLAM IV
Deep VO and SLAM IVYu Huang
 
Depth Fusion from RGB and Depth Sensors III
Depth Fusion from RGB and Depth Sensors  IIIDepth Fusion from RGB and Depth Sensors  III
Depth Fusion from RGB and Depth Sensors IIIYu Huang
 

Was ist angesagt? (20)

Driving behaviors for adas and autonomous driving XII
Driving behaviors for adas and autonomous driving XIIDriving behaviors for adas and autonomous driving XII
Driving behaviors for adas and autonomous driving XII
 
Prediction and planning for self driving at waymo
Prediction and planning for self driving at waymoPrediction and planning for self driving at waymo
Prediction and planning for self driving at waymo
 
Deep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data IIDeep Learning’s Application in Radar Signal Data II
Deep Learning’s Application in Radar Signal Data II
 
Pedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving VPedestrian behavior/intention modeling for autonomous driving V
Pedestrian behavior/intention modeling for autonomous driving V
 
Driving Behavior for ADAS and Autonomous Driving VIII
Driving Behavior for ADAS and Autonomous Driving VIIIDriving Behavior for ADAS and Autonomous Driving VIII
Driving Behavior for ADAS and Autonomous Driving VIII
 
Driving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VIIDriving Behavior for ADAS and Autonomous Driving VII
Driving Behavior for ADAS and Autonomous Driving VII
 
Deep VO and SLAM
Deep VO and SLAMDeep VO and SLAM
Deep VO and SLAM
 
Driving Behavior for ADAS and Autonomous Driving III
Driving Behavior for ADAS and Autonomous Driving IIIDriving Behavior for ADAS and Autonomous Driving III
Driving Behavior for ADAS and Autonomous Driving III
 
Lidar for Autonomous Driving II (via Deep Learning)
Lidar for Autonomous Driving II (via Deep Learning)Lidar for Autonomous Driving II (via Deep Learning)
Lidar for Autonomous Driving II (via Deep Learning)
 
BEV Semantic Segmentation
BEV Semantic SegmentationBEV Semantic Segmentation
BEV Semantic Segmentation
 
Camera-Based Road Lane Detection by Deep Learning II
Camera-Based Road Lane Detection by Deep Learning IICamera-Based Road Lane Detection by Deep Learning II
Camera-Based Road Lane Detection by Deep Learning II
 
Depth Fusion from RGB and Depth Sensors IV
Depth Fusion from RGB and Depth Sensors  IVDepth Fusion from RGB and Depth Sensors  IV
Depth Fusion from RGB and Depth Sensors IV
 
Fisheye-Omnidirectional View in Autonomous Driving III
Fisheye-Omnidirectional View in Autonomous Driving IIIFisheye-Omnidirectional View in Autonomous Driving III
Fisheye-Omnidirectional View in Autonomous Driving III
 
Deep Learning’s Application in Radar Signal Data
Deep Learning’s Application in Radar Signal DataDeep Learning’s Application in Radar Signal Data
Deep Learning’s Application in Radar Signal Data
 
Simulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atgSimulation for autonomous driving at uber atg
Simulation for autonomous driving at uber atg
 
Jointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planningJointly mapping, localization, perception, prediction and planning
Jointly mapping, localization, perception, prediction and planning
 
3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II3-d interpretation from single 2-d image for autonomous driving II
3-d interpretation from single 2-d image for autonomous driving II
 
camera-based Lane detection by deep learning
camera-based Lane detection by deep learningcamera-based Lane detection by deep learning
camera-based Lane detection by deep learning
 
Deep VO and SLAM IV
Deep VO and SLAM IVDeep VO and SLAM IV
Deep VO and SLAM IV
 
Depth Fusion from RGB and Depth Sensors III
Depth Fusion from RGB and Depth Sensors  IIIDepth Fusion from RGB and Depth Sensors  III
Depth Fusion from RGB and Depth Sensors III
 

Ähnlich wie Pedestrian Behavior/Intention Modeling for Autonomous Driving VI

Adaptive Feature Fusion Networks for Origin-Destination Passenger Flow Predic...
Adaptive Feature Fusion Networks for Origin-Destination Passenger Flow Predic...Adaptive Feature Fusion Networks for Origin-Destination Passenger Flow Predic...
Adaptive Feature Fusion Networks for Origin-Destination Passenger Flow Predic...Shakas Technologies
 
IRJET - A Review on Pedestrian Behavior Prediction for Intelligent Transport ...
IRJET - A Review on Pedestrian Behavior Prediction for Intelligent Transport ...IRJET - A Review on Pedestrian Behavior Prediction for Intelligent Transport ...
IRJET - A Review on Pedestrian Behavior Prediction for Intelligent Transport ...IRJET Journal
 
Coupled Layer-wise Graph Convolution for Transportation Demand Prediction
Coupled Layer-wise Graph Convolution for Transportation Demand PredictionCoupled Layer-wise Graph Convolution for Transportation Demand Prediction
Coupled Layer-wise Graph Convolution for Transportation Demand Predictionivaderivader
 
IEEE Camad20 presentation - Isam Al Jawarneh
IEEE Camad20 presentation - Isam Al JawarnehIEEE Camad20 presentation - Isam Al Jawarneh
IEEE Camad20 presentation - Isam Al JawarnehIsam Al Jawarneh, PhD
 
Scalable algorithms for nearest neighbor joins on big trajectory data
Scalable algorithms for nearest neighbor joins on big trajectory dataScalable algorithms for nearest neighbor joins on big trajectory data
Scalable algorithms for nearest neighbor joins on big trajectory dataieeepondy
 
Prediction of nodes mobility in 3-D space
Prediction of nodes mobility in 3-D space Prediction of nodes mobility in 3-D space
Prediction of nodes mobility in 3-D space IJECEIAES
 
Path Planning And Navigation
Path Planning And NavigationPath Planning And Navigation
Path Planning And Navigationguest90654fd
 
Path Planning And Navigation
Path Planning And NavigationPath Planning And Navigation
Path Planning And Navigationguest90654fd
 
Deep Multi-View Spatial-Temporal Network for Taxi Demand Prediction
Deep Multi-View Spatial-Temporal Network for Taxi Demand PredictionDeep Multi-View Spatial-Temporal Network for Taxi Demand Prediction
Deep Multi-View Spatial-Temporal Network for Taxi Demand Predictionivaderivader
 
[20240408_LabSeminar_Huy]PivotalSTGNN.pptx
[20240408_LabSeminar_Huy]PivotalSTGNN.pptx[20240408_LabSeminar_Huy]PivotalSTGNN.pptx
[20240408_LabSeminar_Huy]PivotalSTGNN.pptxthanhdowork
 
Cloudsim t-drive enhancing driving directions with taxi drivers’ intelligence
Cloudsim  t-drive enhancing driving directions with taxi drivers’ intelligenceCloudsim  t-drive enhancing driving directions with taxi drivers’ intelligence
Cloudsim t-drive enhancing driving directions with taxi drivers’ intelligenceecway
 
Java t-drive enhancing driving directions with taxi drivers’ intelligence
Java  t-drive enhancing driving directions with taxi drivers’ intelligenceJava  t-drive enhancing driving directions with taxi drivers’ intelligence
Java t-drive enhancing driving directions with taxi drivers’ intelligenceecwayerode
 
T drive enhancing driving directions with taxi drivers’ intelligence
T drive enhancing driving directions with taxi drivers’ intelligenceT drive enhancing driving directions with taxi drivers’ intelligence
T drive enhancing driving directions with taxi drivers’ intelligenceEcway Technologies
 
Java t-drive enhancing driving directions with taxi drivers’ intelligence
Java  t-drive enhancing driving directions with taxi drivers’ intelligenceJava  t-drive enhancing driving directions with taxi drivers’ intelligence
Java t-drive enhancing driving directions with taxi drivers’ intelligenceEcway Technologies
 
Android t-drive enhancing driving directions with taxi drivers’ intelligence
Android  t-drive enhancing driving directions with taxi drivers’ intelligenceAndroid  t-drive enhancing driving directions with taxi drivers’ intelligence
Android t-drive enhancing driving directions with taxi drivers’ intelligenceecway
 
T drive enhancing driving directions with taxi drivers’ intelligence
T drive enhancing driving directions with taxi drivers’ intelligenceT drive enhancing driving directions with taxi drivers’ intelligence
T drive enhancing driving directions with taxi drivers’ intelligenceEcway Technologies
 
Dotnet t-drive enhancing driving directions with taxi drivers’ intelligence
Dotnet  t-drive enhancing driving directions with taxi drivers’ intelligenceDotnet  t-drive enhancing driving directions with taxi drivers’ intelligence
Dotnet t-drive enhancing driving directions with taxi drivers’ intelligenceEcway Technologies
 
5438-Article Text-8663-1-10-20200511.pdf
5438-Article Text-8663-1-10-20200511.pdf5438-Article Text-8663-1-10-20200511.pdf
5438-Article Text-8663-1-10-20200511.pdfTadiyosHailemichael
 

Ähnlich wie Pedestrian Behavior/Intention Modeling for Autonomous Driving VI (20)

Adaptive Feature Fusion Networks for Origin-Destination Passenger Flow Predic...
Adaptive Feature Fusion Networks for Origin-Destination Passenger Flow Predic...Adaptive Feature Fusion Networks for Origin-Destination Passenger Flow Predic...
Adaptive Feature Fusion Networks for Origin-Destination Passenger Flow Predic...
 
IRJET - A Review on Pedestrian Behavior Prediction for Intelligent Transport ...
IRJET - A Review on Pedestrian Behavior Prediction for Intelligent Transport ...IRJET - A Review on Pedestrian Behavior Prediction for Intelligent Transport ...
IRJET - A Review on Pedestrian Behavior Prediction for Intelligent Transport ...
 
Coupled Layer-wise Graph Convolution for Transportation Demand Prediction
Coupled Layer-wise Graph Convolution for Transportation Demand PredictionCoupled Layer-wise Graph Convolution for Transportation Demand Prediction
Coupled Layer-wise Graph Convolution for Transportation Demand Prediction
 
IEEE Camad20 presentation - Isam Al Jawarneh
IEEE Camad20 presentation - Isam Al JawarnehIEEE Camad20 presentation - Isam Al Jawarneh
IEEE Camad20 presentation - Isam Al Jawarneh
 
2006.11583.pdf
2006.11583.pdf2006.11583.pdf
2006.11583.pdf
 
Scalable algorithms for nearest neighbor joins on big trajectory data
Scalable algorithms for nearest neighbor joins on big trajectory dataScalable algorithms for nearest neighbor joins on big trajectory data
Scalable algorithms for nearest neighbor joins on big trajectory data
 
Where Next
Where NextWhere Next
Where Next
 
Prediction of nodes mobility in 3-D space
Prediction of nodes mobility in 3-D space Prediction of nodes mobility in 3-D space
Prediction of nodes mobility in 3-D space
 
Path Planning And Navigation
Path Planning And NavigationPath Planning And Navigation
Path Planning And Navigation
 
Path Planning And Navigation
Path Planning And NavigationPath Planning And Navigation
Path Planning And Navigation
 
Deep Multi-View Spatial-Temporal Network for Taxi Demand Prediction
Deep Multi-View Spatial-Temporal Network for Taxi Demand PredictionDeep Multi-View Spatial-Temporal Network for Taxi Demand Prediction
Deep Multi-View Spatial-Temporal Network for Taxi Demand Prediction
 
[20240408_LabSeminar_Huy]PivotalSTGNN.pptx
[20240408_LabSeminar_Huy]PivotalSTGNN.pptx[20240408_LabSeminar_Huy]PivotalSTGNN.pptx
[20240408_LabSeminar_Huy]PivotalSTGNN.pptx
 
Cloudsim t-drive enhancing driving directions with taxi drivers’ intelligence
Cloudsim  t-drive enhancing driving directions with taxi drivers’ intelligenceCloudsim  t-drive enhancing driving directions with taxi drivers’ intelligence
Cloudsim t-drive enhancing driving directions with taxi drivers’ intelligence
 
Java t-drive enhancing driving directions with taxi drivers’ intelligence
Java  t-drive enhancing driving directions with taxi drivers’ intelligenceJava  t-drive enhancing driving directions with taxi drivers’ intelligence
Java t-drive enhancing driving directions with taxi drivers’ intelligence
 
T drive enhancing driving directions with taxi drivers’ intelligence
T drive enhancing driving directions with taxi drivers’ intelligenceT drive enhancing driving directions with taxi drivers’ intelligence
T drive enhancing driving directions with taxi drivers’ intelligence
 
Java t-drive enhancing driving directions with taxi drivers’ intelligence
Java  t-drive enhancing driving directions with taxi drivers’ intelligenceJava  t-drive enhancing driving directions with taxi drivers’ intelligence
Java t-drive enhancing driving directions with taxi drivers’ intelligence
 
Android t-drive enhancing driving directions with taxi drivers’ intelligence
Android  t-drive enhancing driving directions with taxi drivers’ intelligenceAndroid  t-drive enhancing driving directions with taxi drivers’ intelligence
Android t-drive enhancing driving directions with taxi drivers’ intelligence
 
T drive enhancing driving directions with taxi drivers’ intelligence
T drive enhancing driving directions with taxi drivers’ intelligenceT drive enhancing driving directions with taxi drivers’ intelligence
T drive enhancing driving directions with taxi drivers’ intelligence
 
Dotnet t-drive enhancing driving directions with taxi drivers’ intelligence
Dotnet  t-drive enhancing driving directions with taxi drivers’ intelligenceDotnet  t-drive enhancing driving directions with taxi drivers’ intelligence
Dotnet t-drive enhancing driving directions with taxi drivers’ intelligence
 
5438-Article Text-8663-1-10-20200511.pdf
5438-Article Text-8663-1-10-20200511.pdf5438-Article Text-8663-1-10-20200511.pdf
5438-Article Text-8663-1-10-20200511.pdf
 

Mehr von Yu Huang

Application of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous DrivingApplication of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous DrivingYu Huang
 
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...Yu Huang
 
Data Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous DrivingData Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous DrivingYu Huang
 
Techniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous DrivingTechniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous DrivingYu Huang
 
BEV Joint Detection and Segmentation
BEV Joint Detection and SegmentationBEV Joint Detection and Segmentation
BEV Joint Detection and SegmentationYu Huang
 
BEV Object Detection and Prediction
BEV Object Detection and PredictionBEV Object Detection and Prediction
BEV Object Detection and PredictionYu Huang
 
Fisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VIFisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VIYu Huang
 
Fisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving VFisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving VYu Huang
 
Fisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IVFisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IVYu Huang
 
Prediction,Planninng & Control at Baidu
Prediction,Planninng & Control at BaiduPrediction,Planninng & Control at Baidu
Prediction,Planninng & Control at BaiduYu Huang
 
Cruise AI under the Hood
Cruise AI under the HoodCruise AI under the Hood
Cruise AI under the HoodYu Huang
 
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)Yu Huang
 
Scenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous DrivingScenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous DrivingYu Huang
 
How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?Yu Huang
 
Annotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous DrivingAnnotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous DrivingYu Huang
 
Multi sensor calibration by deep learning
Multi sensor calibration by deep learningMulti sensor calibration by deep learning
Multi sensor calibration by deep learningYu Huang
 
Data pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous drivingData pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous drivingYu Huang
 
Open Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planningOpen Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planningYu Huang
 
Lidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rainLidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rainYu Huang
 
Autonomous Driving of L3/L4 Commercial trucks
Autonomous Driving of L3/L4 Commercial trucksAutonomous Driving of L3/L4 Commercial trucks
Autonomous Driving of L3/L4 Commercial trucksYu Huang
 

Mehr von Yu Huang (20)

Application of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous DrivingApplication of Foundation Model for Autonomous Driving
Application of Foundation Model for Autonomous Driving
 
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...The New Perception Framework  in Autonomous Driving: An Introduction of BEV N...
The New Perception Framework in Autonomous Driving: An Introduction of BEV N...
 
Data Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous DrivingData Closed Loop in Simulation Test of Autonomous Driving
Data Closed Loop in Simulation Test of Autonomous Driving
 
Techniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous DrivingTechniques and Challenges in Autonomous Driving
Techniques and Challenges in Autonomous Driving
 
BEV Joint Detection and Segmentation
BEV Joint Detection and SegmentationBEV Joint Detection and Segmentation
BEV Joint Detection and Segmentation
 
BEV Object Detection and Prediction
BEV Object Detection and PredictionBEV Object Detection and Prediction
BEV Object Detection and Prediction
 
Fisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VIFisheye based Perception for Autonomous Driving VI
Fisheye based Perception for Autonomous Driving VI
 
Fisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving VFisheye/Omnidirectional View in Autonomous Driving V
Fisheye/Omnidirectional View in Autonomous Driving V
 
Fisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IVFisheye/Omnidirectional View in Autonomous Driving IV
Fisheye/Omnidirectional View in Autonomous Driving IV
 
Prediction,Planninng & Control at Baidu
Prediction,Planninng & Control at BaiduPrediction,Planninng & Control at Baidu
Prediction,Planninng & Control at Baidu
 
Cruise AI under the Hood
Cruise AI under the HoodCruise AI under the Hood
Cruise AI under the Hood
 
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
LiDAR in the Adverse Weather: Dust, Snow, Rain and Fog (2)
 
Scenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous DrivingScenario-Based Development & Testing for Autonomous Driving
Scenario-Based Development & Testing for Autonomous Driving
 
How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?How to Build a Data Closed-loop Platform for Autonomous Driving?
How to Build a Data Closed-loop Platform for Autonomous Driving?
 
Annotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous DrivingAnnotation tools for ADAS & Autonomous Driving
Annotation tools for ADAS & Autonomous Driving
 
Multi sensor calibration by deep learning
Multi sensor calibration by deep learningMulti sensor calibration by deep learning
Multi sensor calibration by deep learning
 
Data pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous drivingData pipeline and data lake for autonomous driving
Data pipeline and data lake for autonomous driving
 
Open Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planningOpen Source codes of trajectory prediction & behavior planning
Open Source codes of trajectory prediction & behavior planning
 
Lidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rainLidar in the adverse weather: dust, fog, snow and rain
Lidar in the adverse weather: dust, fog, snow and rain
 
Autonomous Driving of L3/L4 Commercial trucks
Autonomous Driving of L3/L4 Commercial trucksAutonomous Driving of L3/L4 Commercial trucks
Autonomous Driving of L3/L4 Commercial trucks
 

Kürzlich hochgeladen

SPICE PARK APR2024 ( 6,793 SPICE Models )
SPICE PARK APR2024 ( 6,793 SPICE Models )SPICE PARK APR2024 ( 6,793 SPICE Models )
SPICE PARK APR2024 ( 6,793 SPICE Models )Tsuyoshi Horigome
 
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)Suman Mia
 
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCollege Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCall Girls in Nagpur High Profile
 
IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...
IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...
IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...RajaP95
 
Introduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptxIntroduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptxupamatechverse
 
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Serviceranjana rawat
 
247267395-1-Symmetric-and-distributed-shared-memory-architectures-ppt (1).ppt
247267395-1-Symmetric-and-distributed-shared-memory-architectures-ppt (1).ppt247267395-1-Symmetric-and-distributed-shared-memory-architectures-ppt (1).ppt
247267395-1-Symmetric-and-distributed-shared-memory-architectures-ppt (1).pptssuser5c9d4b1
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxJoão Esperancinha
 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVRajaP95
 
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...ranjana rawat
 
UNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and workingUNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and workingrknatarajan
 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSISrknatarajan
 
Coefficient of Thermal Expansion and their Importance.pptx
Coefficient of Thermal Expansion and their Importance.pptxCoefficient of Thermal Expansion and their Importance.pptx
Coefficient of Thermal Expansion and their Importance.pptxAsutosh Ranjan
 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escortsranjana rawat
 
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur EscortsCall Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
 
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICSHARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICSRajkumarAkumalla
 
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130Suhani Kapoor
 

Kürzlich hochgeladen (20)

Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCRCall Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
 
SPICE PARK APR2024 ( 6,793 SPICE Models )
SPICE PARK APR2024 ( 6,793 SPICE Models )SPICE PARK APR2024 ( 6,793 SPICE Models )
SPICE PARK APR2024 ( 6,793 SPICE Models )
 
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)
 
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service NashikCollege Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
College Call Girls Nashik Nehal 7001305949 Independent Escort Service Nashik
 
IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...
IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...
IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...
 
Introduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptxIntroduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptx
 
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
 
247267395-1-Symmetric-and-distributed-shared-memory-architectures-ppt (1).ppt
247267395-1-Symmetric-and-distributed-shared-memory-architectures-ppt (1).ppt247267395-1-Symmetric-and-distributed-shared-memory-architectures-ppt (1).ppt
247267395-1-Symmetric-and-distributed-shared-memory-architectures-ppt (1).ppt
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
 
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
UNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and workingUNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and working
 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSIS
 
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
 
Roadmap to Membership of RICS - Pathways and Routes
Roadmap to Membership of RICS - Pathways and RoutesRoadmap to Membership of RICS - Pathways and Routes
Roadmap to Membership of RICS - Pathways and Routes
 
Coefficient of Thermal Expansion and their Importance.pptx
Coefficient of Thermal Expansion and their Importance.pptxCoefficient of Thermal Expansion and their Importance.pptx
Coefficient of Thermal Expansion and their Importance.pptx
 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
 
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur EscortsCall Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
 
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICSHARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
 
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
 

Pedestrian Behavior/Intention Modeling for Autonomous Driving VI

  • 1. Pedestrian Behavior/Intention Modeling for Autonomous Driving VI Yu Huang Yu.huang07@gmail.com Sunnyvale, California
  • 2. Outline • CoMoGCN: Coherent Motion Aware Trajectory Prediction with Graph Representation (5.5) • STINet: Spatio-Temporal-Interactive Network for Pedestrian Detect. and Trajectory Pred. • AC-VRNN: Attentive Conditional-VRNN for Multi-Future Trajectory Prediction (5.17) • Spatio-Temporal Graph Transformer Networks for Pedestrian Trajectory Prediction (5.18) • Intention-aware Residual Bidirectional LSTM for Long-term Pedestrian Trajectory Prediction (6.30) • It Is Not the Journey but the Destination: Endpoint Conditioned Trajectory Prediction (7.6) • Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian Trajectory Prediction by Learning Adaptive Neighborhoods (7.8) • Probabilistic Crowd GAN: Multimodal Pedestrian Trajectory Prediction using a Graph Vehicle-Pedestrian Attention Network (7.12)
  • 3. CoMoGCN: Coherent Motion Aware Trajectory Prediction with Graph Representation • Forecasting human trajectories is critical for tasks such as robot crowd navigation and autonomous driving. • Modeling social interactions is of great importance for accurate group-wise motion prediction. • However, most existing methods do not consider information about coherence within the crowd, but rather only pairwise interactions. • A framework, coherent motion aware graph convolutional network (CoMoGCN), for trajectory prediction in crowded scenes with group constraints. • First, cluster pedestrian trajectories into groups according to motion coherence. • Then, use graph convolutional networks to aggregate crowd information efficiently. • The CoMoGCN also takes advantage of variational autoencoders to capture the multimodal nature of the human trajectories by modeling the distribution.
  • 4. CoMoGCN: Coherent Motion Aware Trajectory Prediction with Graph Representation System overview. procedures: 1. obtain coherent motion labels for each human in an offline data pre-processing procedure. 2. Based on the coherent motion labels for each human, establish graphs capturing intergroup and intragroup relationships. The encoder LSTM takes past trajectories as input and feeds the encoded features into two GCNs. 3. The embeddings from the two GCNs are concatenated and forwarded to an MLP to create a distribution. Then, features are sampled from the distribution and fed into a decoder LSTM for trajectory prediction.
  • 5. CoMoGCN: Coherent Motion Aware Trajectory Prediction with Graph Representation
  • 6. STINet: Spatio-Temporal-Interactive Network for Pedestrian Detection and Trajectory Prediction • Detecting pedestrians and predicting future trajectories for them are critical tasks for numerous applications, such as autonomous driving. • Previous methods either treat the detection and prediction as separate tasks or simply add a trajectory regression head on top of a detector. • An end-to-end two-stage network: Spatio-Temporal-Interactive Network (STINet). • In addition to 3D geometry modeling of pedestrians, model the temporal information for each of the pedestrians. • It predicts both current and past locations in the first stage, so that each pedestrian can be linked across frames and the comprehensive spatio-temporal information can be captured in the second stage. • Also, model the interaction among objects with an interaction graph, to gather the information among the neighboring objects. • Comprehensive experiments on the Lyft Dataset and the recently released large-scale Waymo Open Dataset for both object detection and future trajectory prediction.
  • 7. STINet: Spatio-Temporal-Interactive Network for Pedestrian Detection and Trajectory Prediction The overview. It takes a sequence of point clouds as input, detects pedestrians and predicts their future trajectories simultaneously. The point clouds are processed by Pillar Feature Encoding to generate Pillar Features. Then each Pillar Feature is fed into a backbone ResUNet to get backbone features. A Temporal Region Proposal Network (T-RPN) takes backbone features and generated temporal proposal with past and current boxes for each object. Spatio-Temporal-Interactive (STI) Feature Extractor learns features for each temporal proposal which are used for final detection and trajectory prediction.
  • 8. STINet: Spatio-Temporal-Interactive Network for Pedestrian Detection and Trajectory Prediction Backbone. Upper: overview of the backbone. The input point cloud sequence is fed to Voxelization and Point net to generate pseudo images, which are then processed by ResNet U-Net to generate final backbone feature sequence. Lower: detailed design of ResNet U-Net.
  • 9. STINet: Spatio-Temporal-Interactive Network for Pedestrian Detection and Trajectory Prediction Spatial-Temporal-Interactive Feature Extractor (STI- FE): Local geometry, local dynamic and history path features are extracted given a temporal proposal. For local geometry and local dynamics features, the yellow areas are used for feature extraction. Relational reasoning is performed across proposals’ local features to generate interactive features.
  • 10. STINet: Spatio-Temporal-Interactive Network for Pedestrian Detection and Trajectory Prediction
  • 11. AC-VRNN: Attentive Conditional-VRNN for Multi-Future Trajectory Prediction • Anticipating human motion in crowded scenarios is essential for developing intelligent transportation systems, social-aware robots and advanced video-surveillance applications. • An important aspect of such task is represented by the inherently multi-modal nature of human paths which makes socially-acceptable multiple futures when human interactions are involved. • A generative model for multi-future trajectory prediction based on Conditional Variational Recurrent Neural Networks (C-VRNNs). • Conditioning relies on prior belief maps, representing most likely moving directions and forcing the model to consider the collective agents’ motion. • Human interactions are modeled in a structured way with a graph attention mechanism, providing an online attentive hidden state refinement of the recurrent estimation. • Compared to sequence-to-sequence methods, this model operates step- by-step, generating more refined and accurate predictions.
  • 12. AC-VRNN: Attentive Conditional-VRNN for Multi-Future Trajectory Prediction trajectory prediction framework for a single time-step. The overall model is composed of a training module (left) and an inference module (right). The former is composed of a recurrent variational autoencoder conditioned on prior belief maps. The hidden state of the RNN is refined with an attentive module for the next step of recurrence. The latter performs the displacements generation through the prior network on hidden states and makes an online computation of the adjacency matrix which defines connections between pairs of nodes.
  • 13. AC-VRNN: Attentive Conditional-VRNN for Multi-Future Trajectory Prediction Scheme of the attentive hidden state refinement process. The adjacency matrix is an irregular block matrix where each block size is defined by the number of pedestrians in the current scene (a). Belief map during training for one sample using heat similarity-based strategy. The map is centred at t − 1 to display the sampled displacements distribution at t (b).
  • 14. AC-VRNN: Attentive Conditional-VRNN for Multi-Future Trajectory Prediction
  • 15. Spatio-Temporal Graph Transformer Networks for Pedestrian Trajectory Prediction • Understanding crowd motion dynamics is critical to real- world applications, e.g., surveillance systems and autonomous driving. • This is challenging because it requires effectively modeling the socially aware crowd spatial interaction and complex temporal dependencies. • attention is the most important factor for trajectory prediction. • STAR, a Spatio-Temporal grAph tRans- former framework, tackles trajectory prediction by only attention mechanisms. STAR models intra-graph crowd interaction by TGConv, a Transformer-based graph convolution mechanism. • The inter-graph temporal dependencies modeled by separate temporal Transformers. • STAR captures complex spatio-temporal interactions by interleaving between spatial and temporal Transformers. • To calibrate the temporal prediction for the long-lasting effect of disappeared pedestrians, apply a read-writable external memory module, consistently being updated by the temporal Transformer.
  • 16. Spatio-Temporal Graph Transformer Networks for Pedestrian Trajectory Prediction STAR models the crowd as a graph and learns spatio-temporal interaction of the crowd motion by interleaving between a graph-based spatial Transformer and a temporal Transformer
  • 17. Spatio-Temporal Graph Transformer Networks for Pedestrian Trajectory Prediction Temporal Transformer and Spatial Transformer. (a) Temporal Transformer treats each pedestrians independently and extracts the temporal dependencies by Transformer model (h is the embedding of pedestrian positions, Q, K and V are the query, key, value matrix in Transformers). (b) Spatial Transformer models the crowd as a graph, and applies TGConv, a Transformer-based message passing graph convolution, to model the social interactions (mi→j is the message from node i to j represented by Transformer attention)
  • 18. Spatio-Temporal Graph Transformer Networks for Pedestrian Trajectory Prediction Network structure of STAR with application to trajectory prediction. In STAR, trajectory prediction is achieved completely by attention mechanisms. STAR inter- leaves spatial Transformer and temporal Transformer in two encoder blocks to extract spatio-temporal pedestrian dependencies. An external read-writable graph memory module helps to smooth the graph embeddings and improve the consistency of temporal predictions. The prediction at Tobs + 1 is added back to history to predict the pedestrian poses at Tobs + 2.
  • 19. Spatio-Temporal Graph Transformer Networks for Pedestrian Trajectory Prediction
  • 20. Intention-aware Residual Bidirectional LSTM for Long-term Pedestrian Trajectory Prediction • Trajectory prediction is one of the key capabilities for robots to safely navigate and interact with pedestrians. • Critical insights from human intention and behavioral patterns need to be effectively integrated into long-term pedestrian behavior forecasting. • An intention-aware motion prediction framework consists of a Residual Bidirectional LSTM (ReBiL) and a mutable intention filter. • Instead of learning step-wise displacement, learning offset to warp a nominal intention-aware linear prediction, giving residual learning a physical intuition. • The intention filter is inspired by genetic algorithms and particle filtering, where particles mutate intention hypotheses throughout the pedestrian’s motion with ReBiL as the motion model. • Experiments on a publicly available dataset under abnormal intention-changing scenarios.
  • 21. Intention-aware Residual Bidirectional LSTM for Long-term Pedestrian Trajectory Prediction Overview of our motion prediction framework. ReBiL (dashed-line arrow) performs both truncated prediction for particle weight update and long-term prediction at t after mutation. Mutable intention filter takes truncated prediction results to update particle weights, and it implements Sequential Importance Resampling (SIR) and mutation mechanism.
  • 22. Intention-aware Residual Bidirectional LSTM for Long-term Pedestrian Trajectory Prediction
  • 23. It Is Not the Journey but the Destination: Endpoint Conditioned Trajectory Prediction • Human trajectory forecasting with multiple socially interacting agents is of critical importance for autonomous navigation in human environments, e.g., for self-driving cars and social robots. • Predicted Endpoint Conditioned Network (PECNet) for flexible human trajectory prediction. • PECNet infers distant trajectory endpoints to assist in long-range multi- modal trajectory prediction. • A non- local social pooling layer enables PECNet to infer diverse yet socially compliant trajectories. • Additionally, a simple “truncation- trick” for improving few-shot multi- modal trajectory prediction performance. • Code https://karttikeya.github.io/publication/htf/
  • 24. It Is Not the Journey but the Destination: Endpoint Conditioned Trajectory Prediction Architecture of PECNet: PECNet uses past history, along with ground truth endpoint to train a VAE for multi-modal endpoint inference. Ground-truth endpoints are denoted by ⋆ whereas x denote the sampled endpoints Gc. The sampled endpoints condition the social-pooling & predictor networks for multi-agent multi- modal trajectory forecasting. Red connections denote the parts utilized only during training. Shades of the same color denote spatio-temporal neighbours encoded with the block diagonal social mask in social pooling module
  • 25. It Is Not the Journey but the Destination: Endpoint Conditioned Trajectory Prediction
  • 26. Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian Trajectory Prediction by Learning Adaptive Neighborhoods • Pedestrian trajectory prediction is a prominent research track that has advanced towards modelling of crowd social and contextual interactions, with extensive usage of Long Short-Term Memory (LSTM) for temporal representation of walking trajectories. • Existing approaches use virtual neighborhoods as a fixed grid for pooling social states of pedestrians with tuning process that controls how social interactions are being captured. • This entails performance customization to specific scenes but lowers the generalization capability of the approaches. • Grid-LSTM, a recent extension of LSTM, which operates over multidimensional feature inputs. • A perspective to interaction modeling by proposing that pedestrian neighborhoods can become adaptive in design. • Grid-LSTM as an encoder to learn about potential future neighborhoods and their influence on pedestrian motion given the visual and the spatial boundaries. • The experiment results clearly illustrate the generalization of our approach across datasets.
  • 27. Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian Trajectory Prediction by Learning Adaptive Neighborhoods The static neighborhood grid fO segments the scene image into several local regions. The dynamic grid fS takes pedestrians trajectories x1,x2 along with their looking angle to stem their social interactions. The output static grid has few highlighted areas, which indicates future neighborhoods where pedestrians would walk.
  • 28. Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian Trajectory Prediction by Learning Adaptive Neighborhoods Full pipeline of G2K kernel. The SRI network encodes Vislets and positional trajectories for each pedestrian trajectory. Then maps them into social grid mask using NLSTMv. The GNN network discretize static context using NLSTMo into ’Visuospatial’ neighborhoods and stores pedestrian contextual awareness in fO. At the consequent step, SRI takes fO and fS, and maps them into the weighted adjacency matrix. This will generate the edge set ν as means of completing graph at time-step t.
  • 29. Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian Trajectory Prediction by Learning Adaptive Neighborhoods Gated Neighborhood Network pipeline. At the beginning, 2DCONV encodes a static image of the scene and forward the features into NLSTM cell which discretizes the environment into a virtual grid.
  • 30. Graph2Kernel Grid-LSTM: A Multi-Cued Model for Pedestrian Trajectory Prediction by Learning Adaptive Neighborhoods
  • 31. Probabilistic Crowd GAN: Multimodal Pedestrian Trajectory Prediction using a Graph Vehicle-Pedestrian Attention Network • Understanding and predicting the intention of pedestrians is essential to enable autonomous vehicles and mobile robots to navigate crowds. • This problem becomes increasingly complex when we consider the uncertainty and multimodality of pedestrian motion, as well as the implicit interactions between members of a crowd, including any response to a vehicle. • Probabilistic Crowd GAN, extends recent work in trajectory prediction, combining Recurrent Neural Networks (RNNs) with Mixture Density Networks (MDNs) to output probabilistic multimodal predictions, from which likely modal paths are found and used for adversarial training. • use of Graph Vehicle-Pedestrian Attention Network (GVAT), which models social interactions and allows input of a shared vehicle feature, showing that inclusion of this module leads to improved trajectory prediction both with and without the presence of a vehicle. • Through evaluation on various datasets, illustrates how the true multimodal and uncertain nature of crowd interactions can be directly modelled.
  • 32. Probabilistic Crowd GAN: Multimodal Pedestrian Trajectory Prediction using a Graph Vehicle-Pedestrian Attention Network Observed pedestrian trajectories are passed to the Generator’s encoder LSTM, whilst the relative position of all agents, including any vehicle, are passed to the GVAT Pooling module. The Generator outputs a GMM for each agent, from which the MultiPAC module finds the likely modal paths, which are compared to ground truth paths by the Discriminator.
  • 33. Probabilistic Crowd GAN: Multimodal Pedestrian Trajectory Prediction using a Graph Vehicle-Pedestrian Attention Network Node features of agent i (red) in GVAT. The distance from i to the vehicle is appended to each other ped-ped distance input before encoding to account for the impact of the vehicle on i’s relationships within the graph
  • 34. Probabilistic Crowd GAN: Multimodal Pedestrian Trajectory Prediction using a Graph Vehicle-Pedestrian Attention Network