<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>XRCE - Seminars</title>
    <link>http://www.xrce.xerox.com/</link>
    <description/>
    <language>en-GB</language>
    <managingEditor>ezp@xrce.xerox.com ()</managingEditor>
    <pubDate>Tue, 05 Sep 2017 10:44:38 +0000</pubDate>
    <lastBuildDate>Tue, 05 Sep 2017 10:44:38 +0000</lastBuildDate>
    <generator>eZ Components Feed dev (http://ezcomponents.org/docs/tutorials/Feed)</generator>
    <docs>http://www.rssboard.org/rss-specification</docs>
    <atom:link href="http://www.xrce.xerox.com/rss/feed/seminars" rel="self" type="application/rss+xml"/>
    <item>
      <title>Universal, unsupervised, and understandable deep image representations</title>
      <link>http://www.xrce.xerox.com/Our-Research/Seminars/2017/Universal-unsupervised-and-understandable-deep-image-representations</link>
      <description>
&lt;div class="object-right"&gt;&lt;div class="content-view-embeddedmedia"&gt;
	&lt;div class="class-image"&gt;

		&lt;div class="attribute-image"&gt;
			&lt;p&gt;
						  				    		

    
        
    
                                                                                                                                            &lt;img src="http://saturne.xrce.xerox.com/var/siteaccesses/storage/images/media/public-media/images/university-oxford/1703009-1-eng-GB/University-Oxford_medium.jpg" width="200" height="200"  style="border: 0px  ;" alt="University Oxford" title="University Oxford" /&gt;
            
    
    
    			  							&lt;/p&gt;
		&lt;/div&gt;

			&lt;/div&gt;
&lt;/div&gt;

&lt;/div&gt;&lt;p&gt;Monday, June 26th, 11am&lt;/p&gt;&lt;p&gt;Speaker:&amp;nbsp;&lt;a href="http://www.robots.ox.ac.uk/~vedaldi/" target="_self"&gt;Andrea Vedaldi&lt;/a&gt;
, associate professor at&amp;nbsp;&lt;i&gt;University of Oxford&lt;/i&gt;, U.K.&amp;nbsp;&lt;/p&gt;&lt;p&gt;Abstract:&amp;nbsp;Modern deep neural networks have taken computer vision by storm. In this talk, I will demonstrate a few applications of large scale neural nets to problems such as object recognition and text spotting. I will also discuss one of the main difficulties in using these models in new applications, namely the need of providing very large datasets of annotated images. I will show how this problem can be alleviated by the use of synthetic data. Then, I will discuss some of our recent research at the core of this technology, aimed at solving some of these problems at a more fundamental level. I will focus on three areas: unsupervised representations, universal representations, and understanding representations.&amp;nbsp;&lt;/p&gt;</description>
      <author>ezp@xrce.xerox.com (Irene MAXWELL)</author>
      <guid isPermaLink="false">ef42b9f61881aeaa48ccd35a11aa11c2</guid>
      <pubDate>Tue, 13 Jun 2017 09:57:10 +0000</pubDate>
    </item>
    <item>
      <title>Learning novel spatio-temporal representations with and without language</title>
      <link>http://www.xrce.xerox.com/Our-Research/Seminars/2017/Learning-novel-spatio-temporal-representations-with-and-without-language</link>
      <description>
&lt;div class="object-right"&gt;&lt;div class="content-view-embeddedmedia"&gt;
	&lt;div class="class-image"&gt;

		&lt;div class="attribute-image"&gt;
			&lt;p&gt;
						  				    		

    
        
    
                                                                                                                                            &lt;img src="http://saturne.xrce.xerox.com/var/siteaccesses/storage/images/media/public-media/images/univwersity-amsterdam/1703005-1-eng-GB/univwersity-Amsterdam_large.jpg" width="300" height="30"  style="border: 0px  ;" alt="univwersity Amsterdam" title="univwersity Amsterdam" /&gt;
            
    
    
    			  							&lt;/p&gt;
		&lt;/div&gt;

			&lt;/div&gt;
&lt;/div&gt;

&lt;/div&gt;&lt;p&gt;Thursday, June 22nd, 11am&lt;/p&gt;&lt;p&gt;Speaker:&amp;nbsp;&lt;a href="http://www.egavves.com/" target="_self"&gt;Efstratios Gavves&lt;/a&gt;
, assistant professor at&amp;nbsp;&lt;i&gt;University of Amsterdam&lt;/i&gt;, The Netherlands&lt;/p&gt;&lt;p&gt;Abstract: In recent years rethinking video-related tasks and their optimal representations have attracted a significant amount of attention, including but not limited to action and event recognition, visual object tracking and spatio-temporal localization. In this talk, I will present our CVPR 2017 accepted work on spatio-temporal representations with and without the use of language.&lt;/p&gt;&lt;p&gt;I first discuss our latest work on rethinking visual object tracking. The basic assumption by visual trackers in the last 30 years was the presence of a bounding box in the first frame. Rather than specifying the target in the first frame of a video by a bounding box, we propose to track the object based on a natural language specification of the target. Tracking by natural language specification not only allows for a more natural human-machine interaction but also improves tracking results. Most importantly, it allows for novel tracking scenarios, for instance, live-streaming tracking or multiple video simultaneous tracking.&lt;/p&gt;&lt;p&gt;Next, I focus on event detection. By definition, events are a complex combination of actions, which can be non-trivially described textually in significant detail, e.g. as in &lt;a href="http://www.wikihow.com/" target="_self"&gt;WikiHow&lt;/a&gt;
. In this work, we show that such external databases that textually describe events allow for improved representation learning. By casting novel events via their textual description as a mixture of known events, we can retrieve relevant videos even with zero-exemplars at training time.&lt;/p&gt;&lt;p&gt;Last, I present a new self-supervised methodology specialized for video, relying on a novel auxiliary task, called odd-one-out learning. In odd-one-out learning, the machine is asked to identify the unrelated or odd element from a set of otherwise related elements. Adapting this to video, we sample sub-videos with the correct and the wrong (odd) order of frames. Our learning machine is implemented as a multi-stream convolutional neural network, which is learned end-to-end and generalizes to other related tasks such as action recognition.&amp;nbsp;&lt;/p&gt;</description>
      <author>ezp@xrce.xerox.com (Claudia HEYER)</author>
      <guid isPermaLink="false">33ab623aee088aff06ec5c7fdbe6fcbc</guid>
      <pubDate>Tue, 13 Jun 2017 09:08:39 +0000</pubDate>
    </item>
    <item>
      <title>Technical mediation of social cognition: minimalist experimental paradigm of perceptual interactions</title>
      <link>http://www.xrce.xerox.com/Our-Research/Seminars/2017/Technical-mediation-of-social-cognition-minimalist-experimental-paradigm-of-perceptual-interactions</link>
      <description>&lt;p&gt;Monday, June 12th, 11am.&lt;/p&gt;
&lt;div class="object-right"&gt;&lt;div class="content-view-embeddedmedia"&gt;
	&lt;div class="class-image"&gt;

		&lt;div class="attribute-image"&gt;
			&lt;p&gt;
						  				    		

    
        
    
                                                                                                                                            &lt;img src="http://saturne.xrce.xerox.com/var/siteaccesses/storage/images/utc/1683435-1-eng-GB/utc_medium.jpg" width="200" height="70"  style="border: 0px  ;" alt="UTC logo" title="UTC logo" /&gt;
            
    
    
    			  							&lt;/p&gt;
		&lt;/div&gt;

			&lt;/div&gt;
&lt;/div&gt;

&lt;/div&gt;&lt;p&gt;Speaker: Charles Lenay,&amp;nbsp;Université de technologie de Compiègne,&amp;nbsp;France&lt;/p&gt;&lt;p&gt;After presenting the context, issues and ambitions of a technological research in the Humanities as promoted by the COSTECH UTC team, I will give some examples of research in cognitive science. These studies, both scientific, technological and philosophical on perceptual supplementation systems are in line with the so-called &amp;quot;4E&amp;quot; approaches in cognitive sciences (Enactive, Embodied, Embedded, Extended). It is a question of constructing very simple experimental situations to highlight the dynamics of interactions that constitute the skills of social cognition as the &amp;quot;recognition of others&amp;quot;, the &amp;quot;perception of her intentions&amp;quot; or &amp;quot;imitation of her facial expressions&amp;quot;.&lt;/p&gt;</description>
      <author>ezp@xrce.xerox.com (Claudia HEYER)</author>
      <guid isPermaLink="false">74aabe54c8ecabd3963a852c44ddb915</guid>
      <pubDate>Sun, 04 Jun 2017 20:46:51 +0000</pubDate>
    </item>
    <item>
      <title>Evolving dialogue with machines</title>
      <link>http://www.xrce.xerox.com/Our-Research/Seminars/2017/Evolving-dialogue-with-machines</link>
      <description>&lt;p&gt;Thursday, June 1st at 11am.&lt;/p&gt;&lt;p&gt;Speaker:&amp;nbsp;Milica Gasic, lecturer, University of Cambridge, UK.&lt;/p&gt;&lt;p&gt;Abstract:&amp;nbsp;&lt;/p&gt;&lt;p&gt;In the last decade we have witnessed machine learning trigger a revolution in dialogue research. Using a variety of reinforcement and supervised learning methods and innovative architectures, we can now build fully data-driven dialogue systems. These techniques are part of a user-in-the-loop framework, where systems can be deployed quickly, handle speech recognition errors gracefully and learn continuously from interaction with real users.&lt;/p&gt;&lt;p&gt;Methods based on Gaussian processes are particularly effective as they enable good models to be estimated from limited training data. Furthermore, they provide an explicit estimate of uncertainty, which is particularly useful for reinforcement learning. This talk explores the additional steps that are necessary to extend these methods to support adaptation to different dialogue domains and more data efficient learning, an important step for scaling up and building evolving systems.&lt;/p&gt;&lt;p&gt;The final part of the talk will focus on the evolution of the next generation of spoken dialogue systems. These systems will need to operate on large and dynamic domains and, more importantly, be capable of conducting rich and natural interaction. We will present a research roadmap towards this goal. A typical application where such a level of complexity is needed is a mental health application that we will discuss to illustrate the need for such research&lt;/p&gt;</description>
      <author>ezp@xrce.xerox.com (Irene MAXWELL)</author>
      <guid isPermaLink="false">70b0d8511dafa254b6336f8f914bbcb7</guid>
      <pubDate>Sun, 04 Jun 2017 20:44:06 +0000</pubDate>
    </item>
    <item>
      <title>Sensitivity classification for assisting digital sensitivity review of government documents</title>
      <link>http://www.xrce.xerox.com/Our-Research/Seminars/2017/Sensitivity-classification-for-assisting-digital-sensitivity-review-of-government-documents</link>
      <description>&lt;p&gt;Wednesday, May 31st at 11am.&lt;/p&gt;&lt;p&gt;Speaker,&amp;nbsp;&lt;a href="http://www.dcs.gla.ac.uk/~graham/" target="_self"&gt;Graham McDonald&lt;/a&gt;
, doctoral candidate,&amp;nbsp;&lt;i&gt;University of Glasgow&lt;/i&gt;, U.K.&lt;/p&gt;&lt;p&gt;Freedom of Information (FOI) laws legislate that government documents should be opened to the public. However, many government documents contain sensitive information, such as personal information or information that would be likely to damage the international relations of countries if it were to be released. Therefore, sensitive information is exempt from released through FOI and all government documents must be sensitivity reviewed to ensure that any such sensitivities are not released to the public. With the emergence of born-digital government documents in recent decades, traditional (paper-based) sensitivity review processes are no longer viable, and there is a timely need for automatic sensitivity classification techniques to assist the sensitivity review of born-digital documents. In this talk, I will begin by providing an overview of some of the main characteristics of sensitive information and the associated challenges for automatic sensitivity classification, before presenting some of our recent work developing sensitivity classifiers that make use of syntactic evidence from sequences of Parts-of-Speech, and semantic evidence derived from word embeddings, to effectively classify personal information and international relations sensitivities in government documents.&amp;nbsp;&lt;/p&gt;</description>
      <author>ezp@xrce.xerox.com (Irene MAXWELL)</author>
      <guid isPermaLink="false">239347d089116a052d9ce597de81d72a</guid>
      <pubDate>Sun, 04 Jun 2017 20:38:02 +0000</pubDate>
    </item>
    <item>
      <title>Towards joint understanding of images and language</title>
      <link>http://www.xrce.xerox.com/Our-Research/Seminars/2017/Towards-joint-understanding-of-images-and-language</link>
      <description>
&lt;div class="object-right"&gt;&lt;div class="content-view-embeddedmedia"&gt;
	&lt;div class="class-image"&gt;

		&lt;div class="attribute-image"&gt;
			&lt;p&gt;
						  				    		

    
        
    
                                                                                                                                            &lt;img src="http://saturne.xrce.xerox.com/var/siteaccesses/storage/images/media/public-media/images/university-of-illinois-at-urbana-champaign/1681168-1-eng-GB/University-of-Illinois-at-Urbana-Champaign_medium.jpg" width="200" height="59"  style="border: 0px  ;" alt="University-of-Illinois-at-Urbana-Champaign" title="University-of-Illinois-at-Urbana-Champaign" /&gt;
            
    
    
    			  							&lt;/p&gt;
		&lt;/div&gt;

			&lt;/div&gt;
&lt;/div&gt;

&lt;/div&gt;&lt;p&gt;Monday, 29th May at 2:00PM&lt;/p&gt;&lt;p&gt;Speaker:&amp;nbsp;Svetlana Lazebnik, associate professor at&amp;nbsp;&lt;i&gt;University of Illinois at Urbana-Champaign&lt;/i&gt;, Urbana, IL, U.S.A&lt;/p&gt;&lt;p&gt;Abstract:&amp;nbsp;Numerous real-world tasks can benefit from practical systems that can identify objects in scenes based on language and understand language grounded in visual context. This presentation will focus on my group's recent work on developing systems for jointly modeling images and language. I will talk about neural models for learning cross-modal embeddings for text-to-image and image-to-text search, and about the challenging task of grounding or localizing of textual mentions of entities in an image. Finally, I will discuss applications of our models to automatic image description and visual question answering.&lt;/p&gt;</description>
      <author>ezp@xrce.xerox.com (Claudia HEYER)</author>
      <guid isPermaLink="false">b6383408e841118bd7ac36aa7e518b0c</guid>
      <pubDate>Wed, 24 May 2017 09:41:46 +0000</pubDate>
    </item>
    <item>
      <title>Neural belief tracker: data-driven dialogue state tracking using semantically specialised vector spaces</title>
      <link>http://www.xrce.xerox.com/Our-Research/Seminars/2017/Neural-belief-tracker-data-driven-dialogue-state-tracking-using-semantically-specialised-vector-spaces</link>
      <description>
&lt;div class="object-right"&gt;&lt;div class="content-view-embeddedmedia"&gt;
	&lt;div class="class-image"&gt;

		&lt;div class="attribute-image"&gt;
			&lt;p&gt;
						  				    		

    
        
    
                                                                                                                                            &lt;img src="http://saturne.xrce.xerox.com/var/siteaccesses/storage/images/media/public-media/images/university-cambridge-logo/1680946-1-eng-GB/University-Cambridge-logo_medium.jpg" width="200" height="42"  style="border: 0px  ;" alt="University Cambridge logo" title="University Cambridge logo" /&gt;
            
    
    
    			  							&lt;/p&gt;
		&lt;/div&gt;

			&lt;/div&gt;
&lt;/div&gt;

&lt;/div&gt;&lt;p&gt;18&lt;sup&gt;th&lt;/sup&gt; May, 2017&lt;/p&gt;&lt;p&gt;Speaker:&amp;nbsp;&lt;a href="https://sites.google.com/site/nikolamrksic/" target="_blank"&gt;Nikola Mrkšić&lt;/a&gt;
,doctoral candidate, University of Cambridge, Cambridge, U.K.&lt;/p&gt;&lt;p&gt;Abstract: One of the core components of modern spoken dialogue systems is the belief tracker, which estimates the user's goal at every step of the dialogue. However, most current approaches have difficulty scaling to larger, more complex dialogue domains. This is due to their dependency on either: a) Spoken Language Understanding models that require large amounts of annotated training data; or b) hand-crafted lexicons for capturing some of the linguistic variation in users' language. We propose a novel Neural Belief Tracking (NBT) framework which overcomes these problems by building on recent advances in representation learning. NBT models reason over pre-trained, semantically specialised word vectors, learning to compose them into distributed representations of user utterances and dialogue context. Our evaluation on two datasets shows that this approach surpasses past limitations, matching the performance of state-of-the-art models which rely on hand-crafted semantic lexicons and outperforming them when such lexicons are not provided. Finally, we will discuss how the properties of underlying vector spaces impact model performance, and how the fact that the proposed model operates purely over word vectors allows immediate deployment of belief tracking models for other languages.&lt;/p&gt;</description>
      <author>ezp@xrce.xerox.com (Claudia HEYER)</author>
      <guid isPermaLink="false">3e8d0de1d50950cc9f289963ef144e7a</guid>
      <pubDate>Thu, 04 May 2017 12:51:00 +0000</pubDate>
    </item>
    <item>
      <title>E2D2: Episodic exploration for deep deterministic policies for Starcraft </title>
      <link>http://www.xrce.xerox.com/Our-Research/Seminars/2017/E2D2-Episodic-exploration-for-deep-deterministic-policies-for-Starcraft</link>
      <description>
&lt;div class="object-right"&gt;&lt;div class="content-view-embeddedmedia"&gt;
	&lt;div class="class-image"&gt;

		&lt;div class="attribute-image"&gt;
			&lt;p&gt;
						  				    		

    
        
    
                                                                                                                                            &lt;img src="http://saturne.xrce.xerox.com/var/siteaccesses/storage/images/media/public-media/images/fair/1671062-1-eng-GB/FAIR_medium.jpg" width="200" height="39"  style="border: 0px  ;" alt="FAIR logo" title="FAIR logo" /&gt;
            
    
    
    			  							&lt;/p&gt;
		&lt;/div&gt;

			&lt;/div&gt;
&lt;/div&gt;

&lt;/div&gt;&lt;p&gt;28&lt;sup&gt;th&lt;/sup&gt;&amp;nbsp;April, 2017 &amp;nbsp; &amp;nbsp; 11:00 AM&lt;/p&gt;&lt;p&gt;Speaker:&amp;nbsp;&lt;a href="https://research.fb.com/people/synnaeve-gabriel/" target="_self"&gt;Gabriel Synnaeve&lt;/a&gt;
, research scientist at&amp;nbsp;&lt;i&gt;Facebook AI Research&lt;/i&gt;, New-York, NY, U.S.A.&lt;/p&gt;&lt;p&gt;Abstract:&amp;nbsp;We consider scenarios from the real-time strategy game StarCraft as new benchmarks for reinforcement learning algorithms. We propose micromanagement tasks, which present the problem of the short-term, low-level control of army members during a battle. From a reinforcement learning point of view, these scenarios are challenging because the state-action space is very large, and because there is no obvious feature representation for the state-action evaluation function. We describe our approach to tackle the micromanagement scenarios with deep neural network controllers from raw state features given by the game engine. In addition, we present a heuristic reinforcement learning algorithm which combines direct exploration in the policy space and backpropagation. This algorithm allows for the collection of traces for learning using deterministic policies, which appears much more efficient than, for example, ε-greedy exploration. Experiments show that with this algorithm, we successfully learn non-trivial strategies for scenarios with armies of up to 15 agents, where both Q-learning and REINFORCE struggle.&amp;nbsp;&lt;/p&gt;</description>
      <author>ezp@xrce.xerox.com (Claudia HEYER)</author>
      <guid isPermaLink="false">7a6b1ec2285a75d754672b59fd0916a0</guid>
      <pubDate>Thu, 20 Apr 2017 12:20:48 +0000</pubDate>
    </item>
    <item>
      <title>Computer Vision @ Scale</title>
      <link>http://www.xrce.xerox.com/Our-Research/Seminars/2017/Computer-Vision-Scale</link>
      <description>
&lt;div class="object-right"&gt;&lt;div class="content-view-embeddedmedia"&gt;
	&lt;div class="class-image"&gt;

		&lt;div class="attribute-image"&gt;
			&lt;p&gt;
						  				    		

    
        
    
                                                                                                                                            &lt;img src="http://saturne.xrce.xerox.com/var/siteaccesses/storage/images/facebook-ai-research/1671040-1-eng-GB/Facebook-AI-Research_medium.jpg" width="200" height="39"  style="border: 0px  ;" alt="Facebook AI Research logo" title="Facebook AI Research logo" /&gt;
            
    
    
    			  							&lt;/p&gt;
		&lt;/div&gt;

			&lt;/div&gt;
&lt;/div&gt;

&lt;/div&gt;&lt;p&gt;27&lt;sup&gt;th&lt;/sup&gt; April, 2017 &amp;nbsp; &amp;nbsp;11:00 AM&lt;/p&gt;&lt;p&gt;Speaker:&amp;nbsp;&lt;a href="https://research.fb.com/people/paluri-manohar/" target="_self"&gt;Manohar Paluri&lt;/a&gt;
, research lead at&amp;nbsp;&lt;i&gt;Facebook AI Research&lt;/i&gt;, Menlo Park, CA, U.S.A.&lt;/p&gt;&lt;p&gt;Abstract:&amp;nbsp;Over the past 5 years the community has made significant strides in the field of Computer Vision. Thanks to large scale datasets, specialized computing in form of GPUs and many breakthroughs in modeling better ConvNet architectures, Computer Vision systems in the wild at scale are becoming a reality. At Facebook AI Research, we want to embark on the journey of making breakthroughs in the field of AI and using them for the benefit of connecting people and helping remove barriers for communication. In that regard, Computer Vision plays a significant role as the media content coming to Facebook is ever increasing and building models that understand this content is crucial in achieving our mission of connecting everyone. In this talk I will gloss over how we think about problems related to Computer Vision at Facebook and touch various aspects related to supervised, semi-supervised, unsupervised learning. I will jump between various research efforts involving representation learning. I will highlight some large-scale applications that use the technology and talk about limitations of current systems.&amp;nbsp;&lt;/p&gt;</description>
      <author>ezp@xrce.xerox.com (Claudia HEYER)</author>
      <guid isPermaLink="false">152e35c886db6f7712ef04b49caa5379</guid>
      <pubDate>Thu, 20 Apr 2017 12:17:21 +0000</pubDate>
    </item>
    <item>
      <title>Massive online analytics for the Internet of Things</title>
      <link>http://www.xrce.xerox.com/Our-Research/Seminars/2017/Massive-online-analytics-for-the-Internet-of-Things</link>
      <description>
&lt;div class="object-right"&gt;&lt;div class="content-view-embeddedmedia"&gt;
	&lt;div class="class-image"&gt;

		&lt;div class="attribute-image"&gt;
			&lt;p&gt;
						  				    		

    
        
    
                                                                                                                                            &lt;img src="http://saturne.xrce.xerox.com/var/siteaccesses/storage/images/media/public-media/images/telecom_paristech/1613337-1-eng-GB/telecom_parisTech_small.png" width="100" height="100"  style="border: 0px  ;" alt="Telecom ParisTech logo image" title="Telecom ParisTech logo image" /&gt;
            
    
    
    			  							&lt;/p&gt;
		&lt;/div&gt;

			&lt;/div&gt;
&lt;/div&gt;

&lt;/div&gt;&lt;p&gt;
Speaker:&amp;nbsp;&lt;a href="http://albertbifet.com/" target="_blank"&gt;Albert Bifet&lt;/a&gt;
, professor at&amp;nbsp;&lt;i&gt;Telecom-Paristech&lt;/i&gt;, Paris, France&lt;br /&gt;Abstract:&amp;nbsp;Big Data and the Internet of Things (IoT) have the potential to fundamentally shift the way we interact with our surroundings. The challenge of deriving insights from the Internet of Things (IoT) has been recognized as one of the most exciting and key opportunities for both academia and industry. Advanced analysis of big data streams from sensors and devices is bound to become a key area of data mining research as the number of applications requiring such processing increases. Dealing with the evolution over time of such data streams, i.e., with concepts that drift or change completely, is one of the core issues in stream mining. In this talk, I will present an overview of data stream mining, and I will introduce some popular open source tools for data stream mining.&lt;/p&gt;</description>
      <author>ezp@xrce.xerox.com (Claudia HEYER)</author>
      <guid isPermaLink="false">a3984637b4954f0e435bb41730558a78</guid>
      <pubDate>Wed, 08 Mar 2017 15:43:02 +0000</pubDate>
    </item>
  </channel>
</rss>