SwePub
Sök i SwePub databas

  Utökad sökning

Träfflista för sökning "WFRF:(Azizpour Hossein 1985 ) "

Sökning: WFRF:(Azizpour Hossein 1985 )

  • Resultat 1-10 av 55
Sortera/gruppera träfflistan
   
NumreringReferensOmslagsbildHitta
1.
  • Azizpour, Hossein, 1985-, et al. (författare)
  • Factors of Transferability for a Generic ConvNet Representation
  • 2016
  • Ingår i: IEEE Transactions on Pattern Analysis and Machine Intelligence. - : IEEE Computer Society Digital Library. - 0162-8828 .- 1939-3539. ; 38:9, s. 1790-1802
  • Tidskriftsartikel (refereegranskat)abstract
    • Evidence is mounting that Convolutional Networks (ConvNets) are the most effective representation learning method for visual recognition tasks. In the common scenario, a ConvNet is trained on a large labeled dataset (source) and the feed-forward units activation of the trained network, at a certain layer of the network, is used as a generic representation of an input image for a task with relatively smaller training set (target). Recent studies have shown this form of representation transfer to be suitable for a wide range of target visual recognition tasks. This paper introduces and investigates several factors affecting the transferability of such representations. It includes parameters for training of the source ConvNet such as its architecture, distribution of the training data, etc. and also the parameters of feature extraction such as layer of the trained ConvNet, dimensionality reduction, etc. Then, by optimizing these factors, we show that significant improvements can be achieved on various (17) visual recognition tasks. We further show that these visual recognition tasks can be categorically ordered based on their similarity to the source task such that a correlation between the performance of tasks and their similarity to the source task w.r.t. the proposed factors is observed.
  •  
2.
  • Sharif Razavian, Ali, 1985-, et al. (författare)
  • CNN features off-the-shelf : An Astounding Baseline for Recognition
  • 2014
  • Ingår i: Proceedings of CVPR 2014.
  • Konferensbidrag (refereegranskat)abstract
    • Recent results indicate that the generic descriptors extracted from the convolutional neural networks are very powerful. This paper adds to the mounting evidence that this is indeed the case. We report on a series of experiments conducted for different recognition tasks using the publicly available code and model of the OverFeat network which was trained to perform object classification on ILSVRC13. We use features extracted from the OverFeat network as a generic image representation to tackle the diverse range of recognition tasks of object image classification, scene recognition, fine grained recognition, attribute detection and image retrieval applied to a diverse set of datasets. We selected these tasks and datasets as they gradually move further away from the original task and data the OverFeat network was trained to solve. Astonishingly, we report consistent superior results compared to the highly tuned state-of-the-art systems in all the visual classification tasks on various datasets. For instance retrieval it consistently outperforms low memory footprint methods except for sculptures dataset. The results are achieved using a linear SVM classifier (or L2 distance in case of retrieval) applied to a feature representation of size 4096 extracted from a layer in the net. The representations are further modified using simple augmentation techniques e.g. jittering. The results strongly suggest that features obtained from deep learning with convolutional nets should be the primary candidate in most visual recognition tasks.
  •  
3.
  • Sharif Razavian, Ali, 1985-, et al. (författare)
  • Persistent Evidence of Local Image Properties in Generic ConvNets
  • 2015
  • Ingår i: Image Analysis. - Cham : Springer Publishing Company. ; , s. 249-262
  • Konferensbidrag (refereegranskat)abstract
    • Supervised training of a convolutional network for object classification should make explicit any information related to the class of objects and disregard any auxiliary information associated with the capture of the image or thevariation within the object class. Does this happen in practice? Although this seems to pertain to the very final layers in the network, if we look at earlier layers we find that this is not the case. Surprisingly, strong spatial information is implicit. This paper addresses this, in particular, exploiting the image representation at the first fully connected layer,i.e. the global image descriptor which has been recently shown to be most effective in a range of visual recognition tasks. We empirically demonstrate evidences for the finding in the contexts of four different tasks: 2d landmark detection, 2d object keypoints prediction, estimation of the RGB values of input image, and recovery of semantic label of each pixel. We base our investigation on a simple framework with ridge rigression commonly across these tasks,and show results which all support our insight. Such spatial information can be used for computing correspondence of landmarks to a good accuracy, but should potentially be useful for improving the training of the convolutional nets for classification purposes.
  •  
4.
  • Azizpour, Hossein, 1985-, et al. (författare)
  • From Generic to Specific Deep Representations for Visual Recognition
  • 2015
  • Ingår i: Proceedings of CVPR 2015. - : IEEE conference proceedings. - 9781467367592
  • Konferensbidrag (refereegranskat)abstract
    • Evidence is mounting that ConvNets are the best representation learning method for recognition. In the common scenario, a ConvNet is trained on a large labeled dataset and the feed-forward units activation, at a certain layer of the network, is used as a generic representation of an input image. Recent studies have shown this form of representation to be astoundingly effective for a wide range of recognition tasks. This paper thoroughly investigates the transferability of such representations w.r.t. several factors. It includes parameters for training the network such as its architecture and parameters of feature extraction. We further show that different visual recognition tasks can be categorically ordered based on their distance from the source task. We then show interesting results indicating a clear correlation between the performance of tasks and their distance from the source task conditioned on proposed factors. Furthermore, by optimizing these factors, we achieve stateof-the-art performances on 16 visual recognition tasks.
  •  
5.
  • Azizpour, Hossein, 1985-, et al. (författare)
  • Self-tuned Visual Subclass Learning with Shared Samples An Incremental Approach
  • 2013
  • Annan publikation (övrigt vetenskapligt/konstnärligt)abstract
    • Computer vision tasks are traditionally defined and eval-uated using semantic categories. However, it is known to thefield that semantic classes do not necessarily correspondto a unique visual class (e.g. inside and outside of a car).Furthermore, many of the feasible learning techniques athand cannot model a visual class which appears consistentto the human eye. These problems have motivated the useof 1) Unsupervised or supervised clustering as a prepro-cessing step to identify the visual subclasses to be used ina mixture-of-experts learning regime. 2) Felzenszwalb etal. part model and other works model mixture assignmentwith latent variables which is optimized during learning 3)Highly non-linear classifiers which are inherently capableof modelling multi-modal input space but are inefficient atthe test time. In this work, we promote an incremental viewover the recognition of semantic classes with varied appear-ances. We propose an optimization technique which incre-mentally finds maximal visual subclasses in a regularizedrisk minimization framework. Our proposed approach uni-fies the clustering and classification steps in a single algo-rithm. The importance of this approach is its compliancewith the classification via the fact that it does not need toknow about the number of clusters, the representation andsimilarity measures used in pre-processing clustering meth-ods a priori. Following this approach we show both quali-tatively and quantitatively significant results. We show thatthe visual subclasses demonstrate a long tail distribution.Finally, we show that state of the art object detection meth-ods (e.g. DPM) are unable to use the tails of this distri-bution comprising 50% of the training samples. In fact weshow that DPM performance slightly increases on averageby the removal of this half of the data.
  •  
6.
  • Azizpour, Hossein, 1985-, et al. (författare)
  • Spotlight the Negatives : A Generalized Discriminative Latent Model
  • 2015
  • Konferensbidrag (refereegranskat)abstract
    • Discriminative latent variable models (LVM) are frequently applied to various visualrecognition tasks. In these systems the latent (hidden) variables provide a formalism formodeling structured variation of visual features. Conventionally, latent variables are de-fined on the variation of the foreground (positive) class. In this work we augment LVMsto includenegativelatent variables corresponding to the background class. We formalizethe scoring function of such a generalized LVM (GLVM). Then we discuss a frameworkfor learning a model based on the GLVM scoring function. We theoretically showcasehow some of the current visual recognition methods can benefit from this generalization.Finally, we experiment on a generalized form of Deformable Part Models with negativelatent variables and show significant improvements on two different detection tasks.
  •  
7.
  • Azizpour, Hossein, 1985- (författare)
  • Visual Representations and Models: From Latent SVM to Deep Learning
  • 2016
  • Doktorsavhandling (övrigt vetenskapligt/konstnärligt)abstract
    • Two important components of a visual recognition system are representation and model. Both involves the selection and learning of the features that are indicative for recognition and discarding those features that are uninformative. This thesis, in its general form, proposes different techniques within the frameworks of two learning systems for representation and modeling. Namely, latent support vector machines (latent SVMs) and deep learning.First, we propose various approaches to group the positive samples into clusters of visually similar instances. Given a fixed representation, the sampled space of the positive distribution is usually structured. The proposed clustering techniques include a novel similarity measure based on exemplar learning, an approach for using additional annotation, and augmenting latent SVM to automatically find clusters whose members can be reliably distinguished from background class. In another effort, a strongly supervised DPM is suggested to study how these models can benefit from privileged information. The extra information comes in the form of semantic parts annotation (i.e. their presence and location). And they are used to constrain DPMs latent variables during or prior to the optimization of the latent SVM. Its effectiveness is demonstrated on the task of animal detection.Finally, we generalize the formulation of discriminative latent variable models, including DPMs, to incorporate new set of latent variables representing the structure or properties of negative samples. Thus, we term them as negative latent variables. We show this generalization affects state-of-the-art techniques and helps the visual recognition by explicitly searching for counter evidences of an object presence.Following the resurgence of deep networks, in the last works of this thesis we have focused on deep learning in order to produce a generic representation for visual recognition. A Convolutional Network (ConvNet) is trained on a largely annotated image classification dataset called ImageNet with $\sim1.3$ million images. Then, the activations at each layer of the trained ConvNet can be treated as the representation of an input image. We show that such a representation is surprisingly effective for various recognition tasks, making it clearly superior to all the handcrafted features previously used in visual recognition (such as HOG in our first works on DPM). We further investigate the ways that one can improve this representation for a task in mind. We propose various factors involving before or after the training of the representation which can improve the efficacy of the ConvNet representation. These factors are analyzed on 16 datasets from various subfields of visual recognition.
  •  
8.
  • Baldassarre, Federico, et al. (författare)
  • Explainability Techniques for Graph Convolutional Networks
  • 2019
  • Konferensbidrag (refereegranskat)abstract
    • Graph Networks are used to make decisions in potentially complex scenarios but it is usually not obvious how or why they made them. In this work, we study the explainability of Graph Network decisions using two main classes of techniques, gradient-based and decomposition-based, on a toy dataset and a chemistry task. Our study sets the ground for future development as well as application to real-world problems.
  •  
9.
  • Baldassarre, Federico, et al. (författare)
  • Explanation-Based Weakly-Supervised Learning of Visual Relations with Graph Networks
  • 2020
  • Ingår i: Proceedings, Part XXVIII Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020. - Cham : Springer Nature. ; , s. 612-630
  • Konferensbidrag (refereegranskat)abstract
    • Visual relationship detection is fundamental for holistic image understanding. However, the localization and classification of (subject, predicate, object) triplets remain challenging tasks, due to the combinatorial explosion of possible relationships, their long-tailed distribution in natural images, and an expensive annotation process. This paper introduces a novel weakly-supervised method for visual relationship detection that relies on minimal image-level predicate labels. A graph neural network is trained to classify predicates in images from a graph representation of detected objects, implicitly encoding an inductive bias for pairwise relations. We then frame relationship detection as the explanation of such a predicate classifier, i.e. we obtain a complete relation by recovering the subject and object of a predicted predicate. We present results comparable to recent fully- and weakly-supervised methods on three diverse and challenging datasets: HICO-DET for human-object interaction, Visual Relationship Detection for generic object-to-object relations, and UnRel for unusual triplets; demonstrating robustness to non-comprehensive annotations and good few-shot generalization.
  •  
10.
  • Baldassarre, Federico, et al. (författare)
  • GraphQA: Protein Model Quality Assessment using Graph Convolutional Networks
  • 2020
  • Ingår i: Bioinformatics. - : Oxford University Press. - 1367-4803 .- 1367-4811 .- 1460-2059. ; 37:3, s. 360-366
  • Tidskriftsartikel (refereegranskat)abstract
    • MotivationProteins are ubiquitous molecules whose function in biological processes is determined by their 3D structure. Experimental identification of a protein’s structure can be time-consuming, prohibitively expensive, and not always possible. Alternatively, protein folding can be modeled using computational methods, which however are not guaranteed to always produce optimal results.GraphQA is a graph-based method to estimate the quality of protein models, that possesses favorable properties such as representation learning, explicit modeling of both sequential and 3D structure, geometric invariance, and computational efficiency.ResultsGraphQA performs similarly to state-of-the-art methods despite using a relatively low number of input features. In addition, the graph network structure provides an improvement over the architecture used in ProQ4 operating on the same input features. Finally, the individual contributions of GraphQA components are carefully evaluated.Availability and implementationPyTorch implementation, datasets, experiments, and link to an evaluation server are available through this GitHub repository: github.com/baldassarreFe/graphqaSupplementary informationSupplementary material is available at Bioinformatics online.
  •  
Skapa referenser, mejla, bekava och länka
  • Resultat 1-10 av 55
Typ av publikation
konferensbidrag (25)
tidskriftsartikel (14)
doktorsavhandling (7)
annan publikation (5)
forskningsöversikt (3)
bokkapitel (1)
visa fler...
visa färre...
Typ av innehåll
refereegranskat (42)
övrigt vetenskapligt/konstnärligt (13)
Författare/redaktör
Azizpour, Hossein, 1 ... (50)
Vinuesa, Ricardo (13)
Smith, Kevin, 1975- (12)
Schlatter, Philipp (10)
Carlsson, Stefan (8)
Guastoni, Luca (8)
visa fler...
Sullivan, Josephine (5)
Baldassarre, Federic ... (5)
Björkman, Mårten, 19 ... (5)
Ban, Yifang (5)
Maki, Atsuto (4)
Sullivan, Josephine, ... (4)
Azizpour, Hossein, A ... (4)
Nascetti, Andrea (4)
Strand, Fredrik (4)
Leite, Iolanda (3)
Sharif Razavian, Ali ... (3)
Smith, Kevin, Associ ... (3)
Strand, F (2)
Nerini, Francesco Fu ... (2)
Razavian, Ali Sharif (2)
Eivazi, Hamidreza (2)
Gupta, S. (1)
Lindholm, P (1)
Golzar, Farzin (1)
Kragic, Danica, 1971 ... (1)
Menéndez Hurtado (, ... (1)
Eklund, M (1)
Hartman, Johan (1)
Robertson, Stephanie (1)
Ek, Carl Henrik (1)
Carlsson, Stefan, Pr ... (1)
Azizpour, Hossein (1)
Dignum, Virginia, Pr ... (1)
Elofsson, Arne (1)
Jensfelt, Patric, 19 ... (1)
Balaam, Madeline (1)
Mallor, Fermin (1)
Carlssom, Stefan (1)
Arefiyan, Mostafa (1)
Naderi Parizi, Sobha ... (1)
Caputo, Barbara, Ass ... (1)
Sullivan, Josephine, ... (1)
Pirsiavash, Hamed, A ... (1)
Dogan, Fethiye Irmak (1)
Sundblad, Yngve (1)
Leuchowius, Karl-Joh ... (1)
Cremers, Daniel (1)
Horvath, Peter (1)
Balassa, Tamas (1)
visa färre...
Lärosäte
Kungliga Tekniska Högskolan (55)
Karolinska Institutet (3)
Umeå universitet (1)
Stockholms universitet (1)
Språk
Engelska (55)
Forskningsämne (UKÄ/SCB)
Naturvetenskap (30)
Teknik (20)
Medicin och hälsovetenskap (4)
Samhällsvetenskap (1)

År

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy