
AISHA UROOJ KHAN
Graduate Research Student @ CRCV, UCF
Neuschwanstein Castle, Germany - September 2018

ABOUT
Computer Vision Researcher
I am a Ph.D. candidate at the Center for Research in Computer Vision (CRCV) at University of Central Florida. More recently, I have been exploring visual reasoning, multimodal learning and grounding for visual question answering. I am being co-advised by Dr. Mubarak Shah and Dr. Niels Lobo.
Some of my previous research topics include semantic segmentation, weakly/unsupervised representation learning, hand detection and hand-based action recognition.
HAND SEGMENTATION
For many applications in First Person Vision, it is necessary to accurately segment not only hands of the camera wearer but also the hands of others with whom he is interacting. In this project, we take a deep look into hand segmentation and hand-based action recognition.
Aisha Urooj Khan, Ali Borji
Analysis of Hand Segmentation in the Wild. (CVPR 2018) [pdf] [Project]
SEGMENTING SKY PIXELS IN IMAGES
The quality of scene parsing, particularly sky classification, decreases in night time images, images involving varying weather conditions, and scene changes due to seasonal weather. We focus specifically on sky segmentation, the task of determining sky and not-sky pixels. As a result of our efforts, we have seen an improvement of 10-15% in the average MCR compared to the prior methods on SkyFinder dataset. We have also improved from an off-the shelf-model in terms of average mIOU by nearly 35%. Further, we analyze our trained models on images w.r.t two aspects: times of day and weather.
Cecilia La Place*, Aisha Urooj Khan*, Ali Borji
Segmenting Sky Pixels in Images. (WACV 2019) [arxiv]
EGOCENTRIC HEIGHT ESTIMATION
Height estimation in first-person videos can be a useful feature for both soft-biometrics and object tracking. Here, we propose a method of estimating the height of an egocentric camera without any calibration or reference points. We used both traditional computer vision approaches and deep learning in order to determine the visual cues that results in best height estimation. Here, we introduce a framework inspired by two stream networks comprising of two Convolutional Neural Networks, one based on spatial information, and one based on information given by optical flow in a frame. Given an egocentric video as an input to the framework, our model yields a height estimate as an output.
Jessica Finocchiaro, Aisha Urooj Khan, Ali Borji.
Egocentric Height Estimation.(WACV 2017) [pdf]