自定义博客皮肤VIP专享

*博客头图:

格式为PNG、JPG,宽度*高度大于1920*100像素,不超过2MB,主视觉建议放在右侧,请参照线上博客头图

请上传大于1920*100像素的图片!

博客底图:

图片格式为PNG、JPG,不超过1MB,可上下左右平铺至整个背景

栏目图:

图片格式为PNG、JPG,图片宽度*高度为300*38像素,不超过0.5MB

主标题颜色:

RGB颜色,例如:#AFAFAF

Hover:

RGB颜色,例如:#AFAFAF

副标题颜色:

RGB颜色,例如:#AFAFAF

自定义博客皮肤

-+
  • 博客(0)
  • 资源 (53)
  • 收藏
  • 关注

空空如也

智东西-自动驾驶系列课第4课课件-低速自动驾驶专用车的技术挑战与前景-智行者联合创始人李晓飞

智东西-自动驾驶系列课第4课课件-低速自动驾驶专用车的技术挑战与前景-智行者联合创始人李晓飞

2018-07-10

人工智能精选论文(自然语言处理)

Recurrent neural network based language model Word representations: A simple and general method for semi-supervised learning Natural Language Processing (Almost) from Scratch Efficient Estimation of Word Representations in Vector Space 。。。。。。 NEURAL MACHINE TRANSLATION BY JOINTLY LEARNING TO ALIGN AND TRANSLATE Learning to Compose Neural Networks for Question Answering

2018-07-06

人工智能精选论文(强化学习)

Human-level control through deep reinforcement learning Playing Atari with Deep Reinforcement Learning Deep Learning for Detecting Robotic Grasps Mastering the game of Go with deep neural networks and tree search ........

2018-07-06

人工智能精选论文(图像识别与图像处理)

2010-boureau-cvpr-10.pdf 2012-farabet-pami-13.pdf 2013-10.1.1.224.9632.pdf 2013-1310.1531v1.pdf 2013-Ji_TPAMI2013.pdf 2013-wang_iccv13.pdf 2014-DeepFace-Closing-the-Gap-to-Human-Level-Performance.pdf 2014-Oquab_Learning_and_Transferring_2014_CVPR_paper.pdf 2016-1411.4389v4.pdf 2015-Xiangxue_Spatial Pyramid Pooling.pdf 。。。。。。 ....... 2016-1506.01497v3.pdf 图像识别深度学习经典论文r-cnn、faster rcnn、rpn、fcn、transfer learning等等

2018-07-06

2016全球人工智能发展报告·产业与应用篇

时间地理分布2 人工智能企业投资排名9 应用领域分析1 3 人工智能技术分支相关产业分析2 0

2018-07-06

2015年中国人工智能应用市场研究报告

人工智能发展现状分析 人工智能应用现状分析 人工智能前景及市场机会分析

2018-07-06

CVPR-2018论文合集五

CVPR论文合集--SLAM DNN CNN 目标检测与目标识别 视频目标分割 图像分割 自然语言处理 自动驾驶

2018-07-05

CVPR-2018论文合集四

CVPR-2018论文合集--SLAM DNN CNN 目标检测与目标识别 视频目标分割 图像分割 自然语言处理 自动驾驶

2018-07-05

CVPR2018论文合集三

CVPR论文合集--SLAM DNN CNN 目标检测与目标识别 视频目标分割 图像分割 自然语言处理 自动驾驶

2018-07-05

CVPR-2018论文合集二

CVPR论文合集--SLAM DNN CNN 目标检测与目标识别 视频目标分割 图像分割 自然语言处理 自动驾驶

2018-07-04

人机融合智能的哲学思考

人机融合智能是一种新型智能形式,它不同于人的智能、也不同于人工智能,是一种跨物种越属性结合的下一代智能科学体系。如果说真就是Being,善就是Should,美就是Being+Should的融合;假设机就是Being,人就是Should,那么人机就是Being+Should的融合。同时,人机融合智能也是东西方文明的共同结晶体现。

2018-07-04

人工智能芯片

随着行业发展环境的趋好,人工智能芯片企业间的竞争将不断加剧,行业内企业间并购整合与资本运作将日趋频繁,优秀的人工智能芯片企业必须重视对行业市场的研究,特别是对企业发展环境和客户需求趋势变化的深入研究。

2018-07-04

促进新一代人工智能产业发展三年行动计划(2018-2020年)

当前,新一轮科技革命和产业变革正在萌发,大数据的形成、理论算法的革新、计算能力的提升及网络设施的演进驱动人工智能发展进入新阶段,智能化成为技术和产业发展的重要方向。人工智能具有显著的溢出效应,将进一步带动其他技术的进步,推动战略性新兴产业总体突破,正在成为推进供给侧结构性改革的新动能、振兴实体经济的新机遇、建设制造强国和网络强国的新引擎。为落实《新一代人工智能发展规划》,深入实施“中国制造2025”,抓住历史机遇,突破重点领域,促进人工智能产业发展,提升制造业智能化水平,推动人工智能和实体经济深度融合,制订本行动计划。

2018-07-04

智能汽车创新发展战略

汽车产业是国民经济重要的战略性、支柱,与人群 汽车产业是国民经济重要的战略性、支柱,与人群 汽车产业是国民经济重要的战略性、支柱,与人群 众生活密切相关。本世纪以来, 我国汽车产业快速发展众生活密切相关。本世纪以来, 我国汽车产业快速发展众生活密切相关。本世纪以来, 我国汽车产业快速发展产业 规 模稳居世界首位 ,综合实力显著增强。随着汽车普及程度不断 ,综合实力显著增强。随着汽车普及程度不断 ,综合实力显著增强。随着汽车普及程度不断 提 高,我国已快速进入汽车社会。当前新一轮科技 高,我国已快速进入汽车社会。当前新一轮科技 高,我国已快速进入汽车社会。当前新一轮科技 高,我国已快速进入汽车社会。当前新一轮科技 革命和产业变 革蓬勃兴起 ,智 能汽车已成为产业发展的战略方向。,智 能汽车已成为产业发展的战略方向。,智 能汽车已成为产业发展的战略方向。能汽车不仅是解决社会面临的交通安全、道路拥堵源消 能汽车不仅是解决社会面临的交通安全、道路拥堵源消 能汽车不仅是解决社会面临的交通安全、道路拥堵源消 耗、环境污染 等问题 的重要手段,更是深化供给侧结构性改革、 实施创新驱动发展战略、 建成 现代化强国的重要支撑,对不断满 现代化强国的重要支撑,对不断满 足人民日益增长的美好生活需要具有十分重意义。为加快 足人民日益增长的美好生活需要具有十分重意义。

2018-07-04

CVPR2018论文合集一

SLAM DNN CNN 目标检测与目标识别 视频目标分割 图像分割 自然语言处理 自动驾驶

2018-07-04

如何基于NeuroPilot 平台打造手机AI

1.NeuroPilot 可实现的手机AI 应用 2.手机AI的挑战 3.NeuroPilot如何克服这些挑战 4.功耗功耗

2018-07-04

如何打造可大规模量产的自主代客泊车系统

1. 如何打造可大规模量产的自主代客泊车系统 2. 纵目科技 3. 智东西

2018-07-04

CodeSLAM — Learning a Compact, Optimisable Representation for Dense Visual SLAM

We present a new compact but dense representation of scene geometry which is conditioned on the intensity data from a single image and generated from a code consisting of a small number of parameters. We are inspired by work both on learned depth from images, and auto-encoders. Our approach is suitable for use in a keyframe-based monocular dense SLAM system: While each keyframe with a code can produce a depth map, the code can be optimised efficiently jointly with pose variables and together with the codes of overlapping keyframes to attain global consistency. Condi- tioning the depth map on the image allows the code to only represent aspects of the local geometry which cannot di- rectly be predicted from the image. We explain how to learn our code representation, and demonstrate its advantageous properties in monocular SLAM.

2018-07-04

Avatar-Net: Multi-scale Zero-shot Style Transfer by Feature Decoration

Zero-shot artistic style transfer is an important image synthesis problem aiming at transferring arbitrary style into content images. However, the trade-off between the gener- alization and efficiency in existing methods impedes a high quality zero-shot style transfer in real-time. In this pa- per, we resolve this dilemma and propose an efficient yet effective Avatar-Net that enables visually plausible multi- scale transfer for arbitrary style. The key ingredient of our method is a style decorator that makes up the content fea- tures by semantically aligned style features from an arbi- trary style image, which does not only holistically match their feature distributions but also preserve detailed style patterns in the decorated features. By embedding this mod- ule into an image reconstruction network that fuses multi- scale style abstractions, the Avatar-Net renders multi-scale stylization for any style image in one feed-forward pass. We demonstrate the state-of-the-art effectiveness and effi- ciency of the proposed method in generating high-quality stylized images, with a series of successive applications in- clude multiple style integration, video stylization and etc.

2018-07-04

Tencent- CNN in MRF: Video Object Segmentation Spatio-Temporal MRF

This paper addresses the problem of video object segmentation, where the initial object mask is given in the first frame of an input video. We propose a novel spatiotemporal Markov Random Field (MRF) model defined over pixels to handle this problem. Unlike conventional MRF models, the spatial dependencies among pixels in our model are encoded by a Convolutional Neural Network (CNN). Specifically, for a given object, the probability of a labeling to a set of spatially neighboring pixels can be predicted by a CNN trained for this specific object. As a result, higher-order, richer dependencies among pixels in the set can be implicitly modeled by the CNN. With temporal dependencies established by optical flow, the resulting MRF model combines both spatial and temporal cues for tackling video object segmentation. However, performing inference in the MRF model is very difficult due to the very highorder dependencies. To this end, we propose a novel CNNembedded algorithm to perform approximate inference in the MRF. This algorithm proceeds by alternating between a temporal fusion step and a feed-forward CNN step. When initialized with an appearance-based one-shot segmentation CNN, our model outperforms the winning entries of the DAVIS 2017 Challenge, without resorting to model ensembling or any dedicated detectors.

2018-07-04

A Twofold Siamese Network for Real-Time Object Tracking

Observing that Semantic features learned in an image classification task and Appearance features learned in a similarity matching task complement each other, we build a twofold Siamese network, named SA-Siam, for real-time object tracking. SA-Siam is composed of a semantic branch and an appearance branch. Each branch is a similarity- learning Siamese network. An important design choice in SA-Siam is to separately train the two branches to keep the heterogeneity of the two types of features. In addi- tion, we propose a channel attention mechanism for the semantic branch. Channel-wise weights are computed ac- cording to the channel activations around the target posi- tion. While the inherited architecture from SiamFC [3] al- lows our tracker to operate beyond real-time, the twofold design and the attention mechanism significantly improve the tracking performance. The proposed SA-Siam outper- forms all other real-time trackers by a large margin on OTB-2013/50/100 benchmarks. 1. Introduction

2018-07-04

A Face-to-Face Neural Conversation Model

Neural networks have recently become good at engaging in dialog. However, current approaches are based solely on verbal text, lacking the richness of a real face-to-face conversation. We propose a neural conversation model that aims to read and generate facial gestures alongside with text. This allows our model to adapt its response based on the “mood” of the conversation. In particular, we introduce an RNN encoder-decoder that exploits the movement of facial muscles, as well as the verbal conversation. The decoder consists of two layers, where the lower layer aims at generating the verbal response and coarse facial expressions, while the second layer fills in the subtle gestures, making the generated output more smooth and natural. We train our neural network by having it “watch” 250 movies. We showcase our joint face-text model in generating more natural conversations through automatic metrics and a human study. We demonstrate an example application with a face-to-face chatting avatar.

2018-07-04

3D-RCNN: Instance-level 3D Object Reconstruction via Render-and-Compare

We present a fast inverse-graphics framework for instance-level 3D scene understanding. We train a deep convolutional network that learns to map image regions to the full 3D shape and pose of all object instances in the image. Our method produces a compact 3D representation of the scene, which can be readily used for applications like autonomous driving. Many traditional 2D vision outputs, like instance segmentations and depth-maps, can be obtained by simply rendering our output 3D scene model. We exploit class-specific shape priors by learning a low dimensional shape-space from collections of CAD models. We present novel representations of shape and pose, that strive towards better 3D equivariance and generalization. In order to exploit rich supervisory signals in the form of 2D annotations like segmentation, we propose a differentiable Render-and-Compare loss that allows 3D shape and pose to be learned with 2D supervision. We evaluate our method on the challenging real-world datasets of Pascal3D+ and KITTI, where we achieve state-of-the-art results.

2018-07-04

3D Semantic Segmentation with Submanifold Sparse Convolutional Networks

Convolutional networks are the de-facto standard for analyzing spatio-temporal data such as images, videos, and 3D shapes. Whilst some of this data is naturally dense (e.g., photos), many other data sources are inherently sparse. Examples include 3D point clouds that were obtained using a LiDAR scanner or RGB-D camera. Standard “dense” implementations of convolutional networks are very inefficient when applied on such sparse data. We introduce new sparse convolutional operations that are designed to process spatially-sparse data more efficiently, and use them to develop spatially-sparse convolutional networks. We demonstrate the strong performance of the resulting models, called submanifold sparse convolutional networks (SSCNs), on two tasks involving semantic segmentation of 3D point clouds. In particular, our models outperform all prior state-of-the-art on the test set of a recent semantic segmentation competition

2018-07-04

3D Pose Estimation and 3D Model Retrieval for Objects in the Wild

We propose a scalable, efficient and accurate approach to retrieve 3D models for objects in the wild. Our contri- bution is twofold. We first present a 3D pose estimation approach for object categories which significantly outper- forms the state-of-the-art on Pascal3D+. Second, we use the estimated pose as a prior to retrieve 3D models which accurately represent the geometry of objects in RGB im- ages. For this purpose, we render depth images from 3D models under our predicted pose and match learned im- age descriptors of RGB images against those of rendered depth images using a CNN-based multi-view metric learn- ing approach. In this way, we are the first to report quanti- tative results for 3D model retrieval on Pascal3D+, where our method chooses the same models as human annota- tors for 50% of the validation images on average. In ad- dition, we show that our method, which was trained purely on Pascal3D+, retrieves rich and accurate 3D models from ShapeNet given RGB images of objects in the wild.

2018-07-04

智能网联汽车评价体系介绍

智能网联汽车评价体系介绍 自动驾驶 智能网联汽车 评价体系 人工之鞥

2018-07-19

20170926-国家智能网联汽车(上海)试点示范区.pdf

20170926-国家智能网联汽车(上海)试点示范区.pdf 自动驾驶 人工智能

2018-07-19

20171024第19届亚太汽车工程年会

20171024第19届亚太汽车工程年会 自动驾驶 车联网 V2X 人工智能 深度学习 新能源

2018-07-16

开放合作,构建自动驾驶新生态

开放合作,构建自动驾驶新生态 Opening and Cooperation to build up a New Eco-system of Autonomous Driving

2018-07-16

VALSE大会现场PPT

VALSE大会现场PPT 深度学习 计算机视觉 图像处理 人工智能 自动驾驶 自然语言处理 SLAM

2018-07-13

CVPR-2018论文合集八

CVPR-2018论文合集--SLAM DNN CNN 目标检测与目标识别 视频目标分割 图像分割 自然语言处理 自动驾驶

2018-07-13

CVPR-2018论文合集六

CVPR-2018论文合集--SLAM DNN CNN 目标检测与目标识别 视频目标分割 图像分割 自然语言处理 自动驾驶

2018-07-13

CVPR-2018论文合集七

CVPR-2018论文合集--SLAM DNN CNN 目标检测与目标识别 视频目标分割 图像分割 自然语言处理 自动驾驶

2018-07-13

智能驾驶进程与毫米波雷达技术动态

智能驾驶进程与毫米波雷达技术动态 自动驾驶感知与传感器技术 毫米波障碍物与行人检测

2018-07-12

第六次ADAS国标制定系列工作会议-第一次会议-2018.1-杭州

第六次ADAS国标制定系列工作会议-第一次会议-2018.1-杭州

2018-07-11

智东西-自动驾驶系列课第5课课件-激光雷达在辅助驾驶和无人驾驶中的应用-镭神智能董事长胡小波

智东西-自动驾驶系列课第5课课件-激光雷达在辅助驾驶和无人驾驶中的应用-镭神智能董事长胡小波

2018-07-09

智东西-自动驾驶系列课第6课课件-新造车务实派如何落地自动驾驶-小鹏汽车肖志光-

智东西-自动驾驶系列课第6课课件-新造车务实派如何落地自动驾驶-小鹏汽车肖志光-

2018-07-09

智东西-自动驾驶系列课第7课课件-基于激光雷达的地图创建与定位-速腾聚创创始人邱纯鑫

智东西-自动驾驶系列课第7课课件-基于激光雷达的地图创建与定位-速腾聚创创始人邱纯鑫

2018-07-09

智东西公开课-自动驾驶系列课课件-李文俊-禾多科技-0522

智东西公开课-自动驾驶系列课课件-李文俊-禾多科技-0522 自动驾驶、代客泊车、深度学习、公开课

2018-07-09

自动驾驶的感知定位与高精地图解决方案-DeepMotion联合创始人

自动驾驶的感知定位与高精地图解决方案-DeepMotion联合创始人-蔡锐 深度学习、SLAM、激光雷达、高精度地图

2018-07-09

空空如也

TA创建的收藏夹 TA关注的收藏夹

TA关注的人

提示
确定要删除当前文章?
取消 删除