Projects
Last Projects
Project Title:
Personalized Gan Based Editing
Students:
Elias Habib and Adi Hanna
Supervisors:
Description:
A significant challenge in GAN-based human face manipulation is maintaining a consistent identity, as existing methods often cause changes in facial characteristics. This work presents a solution that integrates personalized generative capabilities with precise image manipulation techniques to address this issue. Our method utilizes a personalized deep generative prior to fine-tune a pre-trained face generator using a small set of portrait images (~100) of an individual, creating a local, low-dimensional manifold in the latent space. This enables semantic editing faithful to the individual’s key facial characteristics while using a user-friendly GUI with a point-to-point dragging method. We additionally enhance the accuracy and stability of image manipulation by employing classical methods such as normalizing latent vectors and expanding our manifold by using known editing direction vectors all that while maintaining comparable runtime efficiency.
Project Title:
Students:
Mais khoury and Christeen Shaheen.
Supervisors:
Description:
This project introduces a novel method for monocular 3D reconstruction, utilizing a deep learning model that leverages Gaussian Splatting in a unique way. Instead of traditional rendering, we use Gaussian Splatting as an efficient 3D latent representation for features within an autoencoder architecture. This approach prioritizes speed and minimal input data, enabling real-time applications on local devices.
Our pipeline defines a 3D latent representation that allows for memory and time-efficient rendering. By learning Gaussian parameters during training, we streamline the reconstruction process, bypassing the need for extensive input data and real-time computation. Extensive evaluations and ablations demonstrate that our model achieves robust and accurate 3D reconstructions from a single image input. This innovation addresses key challenges in the field, significantly enhancing reconstruction efficiency and speed without compromising accuracy.
Project Title:
Prediction of B/T Subtype, and ETV6-RUNX1 Translocation in Pediatric Acute Lymphoblastic Leukemia By Analysis of Giemsa-Stained Whole Slide Images
Students:
Supervisors:
Gil Shamai and Roy Velich
Description:
In this project, we analyze digitized Giemsa-stained bone marrow samples using deep learning techniques to predict B/T subtype and ETV6-RUNX1 translocation in Acute Lymphoblastic Leukemia (ALL). We developed a solution to predict patient-specific medical properties from digitized slides in a statistically significant manner, employing a convolutional neural network (CNN) trained through supervised learning on labeled medical data provided by multiple institutions. Moreover, we experimented with attention-based methods to further improve our results. Our results show that it is possible to predict these medical characteristics using a CNN-based method, but we need more data to further exploit the attention-based method.
Project Title:
Cannabis Detection by Microscope Inspection
Students:
Emma Attal and Arthur Soussan
Supervisors:
Alon Zvirin and Yaron Honen
Description:
Synthetic cannabinoids contain dangerous chemicals that may be much more dangerous than regular cannabis, and its production has increased recently. Therefore, the police are interested in identifying them and differentiate between real and synthetic cannabis. The forensic department of the Israel police collaborated with the Geometric Image Processing (GIP) laboratory in the Technion to create an improved identification using innovative technology tools. Our project developed computer vision based models to identify real cannabis from microscopic pictures of cannabis.
Project Title:
Computational oncology by analysis and alignment of histopathology slide images
Students:
Neta Becker and Shelly Francis
Supervisors:
Roy Velich and Tal Neoran and Gil Shamai
Description:
In modern medicine, biopsy is a common method for diagnosing breast cancer. The tissue sample is sliced into ultra-thin sections, each of which is stained using techniques such as IHC or H&E to highlight different features. A pathologist then examines the stained sections under a microscope to determine if cancerous cells are present.
While IHC staining provides valuable information, it is expensive and time-consuming. H&E staining, on the other hand, is quicker and more cost-effective. Our goal was to reduce the reliance on IHC staining by calculating the alignment between IHC and H&E stained slides. This would allow us to train neural networks using both types of slides, transfer pathological information from one slide to another, and potentially rely solely on H&E stained slides.
To achieve this goal, we developed a GUI that allows matching points between pairs of slides. After obtaining a set of matched points for each pair, we explored various methods for calculating the alignment, including automatic alignment, manual alignment, and triangulation. We examined different models based on various articles and proposed methods for comparing the results achieved through each method in order to develop an algorithm for our goal
Project Title:
Movement Analysis In Heart Tissue
Students:
David Berestetsky , Elad Tsur and Ofir Eldar
Supervisors:
Noam Rotstein and Idan Haim and Dr. Oren Caspi
Description:
Stem cells are being cultivated in Dr. Oren Caspi's medical research laboratory at the faculty of medicine. These stem cells are specifically categorized as cardiomyocytes. Using these cardiomyocytes, 3D tissues are constructed around two silicon rectangles. The movement of the rectangles is driven by the pulses generated by the tissues. It is established that there exists a direct correlation between the displacement of the columns and the level of force exerted by the tissue.
Optimization and Automation of Movement recognition on pre-captured lab videos.
The program will analyze the active force of the captured tissue, by recognizing the movement of rectangular-silicone implants inside of it.
Project Title:
Students:
Lior Grauer and Eden Konopnicki
Supervisors:
Tal Neoran and Roy Velich
Description:
In our project we implemented a domain adaptation model, Domain adaptation refers to the process of adapting a machine learning model trained on a source domain to perform effectively on a target domain. The ultimate goal of domain adaptation is to enable models to generalize well across different domains, allowing them to perform accurately and reliably in real-world settings.
Project Title:
Nested Diffusion Processes for Anytime Image Generation
Students:
Supervisors:
Bahjat Kawar & Michael Elad
Description:
In this work,
we propose an anytime diffusion-based method
that can generate viable images when stopped at
arbitrary times before completion. Using existing
pretrained diffusion models, we show that the generation
scheme can be recomposed as two nested
diffusion processes, enabling fast iterative refinement
of a generated image. We use this Nested
Diffusion approach to peek into the generation
process and enable flexible scheduling based on
the instantaneous preference of the user. In experiments
on ImageNet and Stable Diffusion-based
text-to-image generation, we show, both qualitatively
and quantitatively, that our method’s intermediate
generation quality greatly exceeds that of
the original diffusion model, while the final slow
generation result remains comparable.
Project Title:
Self-Supervised Learning Methods amp; Vision Transformer Architecture for extracting global and meaningful representations
Students:
Supervisors:
Description:
In our project, we implemented and deployed SSL architecture called Dino, a self-distillation, non-contrastive learning approach. This architecture is model agnostic but empirically shown to work best with Vision Transformer (ViT) architecture.
The models that were trained using the SSL architecture enabled us to assess whether this kind of SSL does indeed produce high-quality features and learn to find underlying structures in the data and identify global relations.
In this report, we will present an overview of previous research in computational histopathology, detail the methods we utilized, examine the experiments we conducted along with their outcomes, and conclude by identifying potential constraints in our proposed solutions and recommending avenues for further improvement or future research.
Project Title:
Deep Learning Approach for Tissue Segmentation in Whole Slide Images
Students:
Haneen Naaran, Amir Bishara
Supervisors:
Description:
We investigate the use of deep learning models for the semantic segmentation of Hematoxylin and Eosin Whole Slide images (H&E WSI). H&E images are commonly used in pathology for the diagnosis.
Automated segmentation can aid in the efficient identification of different structures within these wsi images. This segmentation problem can be solved by classic computer vision algorithms, like Otsu’s thresholding method. However, this method does not generalize well on different datasets and is not robust to artifacts on scans. Hence, we studied the use of deep learning models for this segmentation task, aiming to get better generalization ability and robustness to artifacts. We trained and tested several deep learning models, including U-Net and FusionNet, in a supervised manner on datasets of H&E images that include foreground tissue segmentation performed by the OTSU’s thresholding method. We evaluated their performance using various metrics, including visual examination.
Our results show that these models can achieve high accuracy in the semantic segmentation and they generalize well on different datasets.
Project Title:
Procedurally generated Space
Students:
Supervisors:
Yaron Honen & Boaz Sterenfeld
Description:
Project's Description: The objective of the project was to create planets using simplex noise, with a focus on developing a level-of-detail (LOD) system that allows for rendering small details while maintaining optimal performance.
Next, create a system that generates an endless space, with planets we made before serving as the building blocks.
Afterward, create a planet-naming system and a feature for locating lost planets within the vast expanse of space.
Project Title:
VR-Hight dynamic avatars scene
Students:
Ran Marashi, Michael Radzeiovsky, Eitan Levi
Supervisors:
Boaz Sternfeld and Yaron Honen
Description:
Realistic life experience in a Japanese village up close using VR with famous characters.
https://ranmarashi7.wixsite.com/vr-project
Project Title:
Students:
Ron Benchetrit and Mor Lavon
Supervisors:
Alon Zvirin and Yaron Honen
Description:
We extend the MSCG-Net with agricultural indices and evaluate the results. We were able to
improve the model results with the addition of only a few parameters.
We later on find that maybe the improvement stems from the extra parameters of the
network and not the indices transforms themself.
Project Title:
High Perceptual Quality Single Image Super Resolution
Students:
David-Elone Zana and Odelia Bellaiche Bensegnor
Supervisors:
Description:
Nowadays, the metric used to calculate the statistic distance between different pictures is the FID (Fréchet Inception Distance): it uses a pretrained inception network and the W2 divergence to approach the distance between them. We assume that the latent representation of each dataset has a Gaussian distribution. We also assume that the Gaussian distribution is not degenerate: we assume that the covariance matrix is a positive definite matrix: all the principal components are not zeros. To these assumptions, we can add the numerical instability and the impossibility to score a single image.
The limits of this method had woke up in us the will to look for a new metric. Can we find a new solution that won’t need those assumptions and with whom we can score a single image while maintaining numerical stability?
Project Title:
Students:
Noofar Ophir and Dor Brekhman
Supervisors:
Yaron Honen & Boaz Sterenfeld
Description:
The goal of the project is to simulate a practice environment for squash players. The scene is a squash court in which the player can grab the ball and hit it with a racket. During the game, the player can practice his squash skills and his counter strikes to a random canon’s serve.
Project Title:
Students:
Daniel Cohen and Meir Friedmann
Supervisors:
Yaron Honen & Boaz Sterenfeld
Description:
Brick Smash VR is a VR game inspired by the widely popular classic video games Breakout and Arkanoid. Our goal was to take these classics and transform them into an immersive VR experience.
In Brick Smash VR, the player plays the part of the paddle under the bricks in the original games (as if looking up at the bricks needed to be broken). The player has to break all the bricks before time runs out or they run out of balls. They start by throwing the ball, and then the same ball can be caught with either of the in-game hands and rethrown, or can be hit with a racket held in either hand. Furthermore, the player tries to collect points along the way by catching the fragments of bricks they destroy. The game also features several powerups that manipulate the game in several ways (moving the brick walls, gaining extra balls, point multipliers, etc.).
Project Title:
Students:
Yuval Reshef, Tal Haklay and Asaf-Yosef Buchnick
Supervisors:
Yaron Honen & Boaz Sterenfeld
Description:
We implement the method proposed in “Photo Wake-Up: 3D Character Animation from a Single Photo” (Weng Et al.). Our pipeline takes a single photo of a human character as an input, and outputs an animated version of it, using a specific animation we decided to focus on. Our project's main contribution is using new and improved libraries to implement the pipeline described in the paper. Website - https://talhaklay535.wixsite.com/photo-wakeup Code - https://github.com/abuchnick/PhotoWakeUp
Project Title:
Students:
David Koplevatzkii and Artem Shtefan
Supervisors:
Yaron Honen & Boaz Sterenfeld
Description:
The goal of the project was to reconstruct the famous scene from "the Matrix" movie, where the main character sees everything in digital rain. Considering previous work - "digital rain" shader applied on a pre-defined 3D scene, our goal was to implement this shader application in real-time. We achieved that using triplanar shader mapped on the LiDAR sensor generated mesh surface.
Project Title:
Students:
Gal Katz, Tamir Daniel and Ari Amitai
Supervisors:
Yaron Honen & Boaz Sterenfeld
Description:
The project is inspired by our childhood game, Chicken Invaders.
We used AR technology to bring Chicken Invaders to life.
Using Microsoft HoloLens & Unity AR technology we succeeded in our mission.
Using Hololens AR, the game is played by scanning the environment and opening portals on walls around the player. Chickens emerge from the portals and shoot eggs at the player. The player can kill the chickens by shooting a laser at them. Once it has been destroyed, it turns into baked chicken just like in the original game. Getting as many chickens as possible is the goal.
Our engine identifies the environment, hand gestures and creates portals with noticeable depth.
We also developed the game for the Android platform using Unity AR Foundation.
Project Title:
Students:
Anat Veikherman and Tal Amir
Supervisors:
Description:
In this project, we build an application that uses a web camera to record the participants' faces and detect their emotions in real-time. The application can detect several faces at once and recognize their emotions. We show the detected face and the emotion on the screen while the face is changing in real-time.
Project Title:
VR Hand Interaction Multiplayer
Students:
Leonid Shleifer and Yotam Portal
Supervisors:
Yaron Honen & Boaz Sterenfeld
Description:
In this project we built a Virtual Reality application that enables two users to enter the same VR shared space and a mutual physical room and then together observe an object, move it around and scale it.
In the application the users can grab the objects and move them in all axes, spin and tilt the objects in every direction, enlarge and diminish the objects, all that while the other user will see in real time the object’s motion and also the first user’s hands representations as two spheres.
The application is built as a VR game in Unity and runs on (Oculus) Meta Quest 2 VR headset, the application is not cross platform since it requires the Meta Quest 2 hand tracking capabilities, more on that later.
A good use for this type of application can be a design team working on a certain product, entering the application and being able to see the design in 3D, and discuss it as if it was in front of them in real life.
Project Title:
Find Closest Defibrillator
Students:
Shadi Abu-Saleh and Sameer Hamada
Supervisors:
Yaron Honen & Boaz Sterenfeld
Description:
we introduce a new mobile App that navigate user to the closest Defibrillator in the Technion campus.
Using our App the user would get navigation assistance to the closest defibrillator based on his current location.
The app provides graphics directions on the user`s camera along with his current location and a 2D map that shows his location and the path to the chosen Defibrillator to his Location.
In addition to the 2D and 3D assistance, Voice Assistance is also available which notify the user to the next direction until he reaches the Defibrillator.
Project Title:
Further Steps in Precise Shape Completion
Students:
Hadas Romov and Yiftach Edelstain
Supervisors:
Description:
In recent reality, people are looking for new ways to connect with each other. Virtual reality can be the new meeting ground for people all around the world. However, capturing a person to view in VR is limited by equipment and capture location. To enable those new applications and usages, we want to achieve a simple setup to enable everyone to participate in this new frontier.
In our project, we have extended the paper "Towards Precise Completion of Deformable Shapes", which uses a prior based approach for shape reconstruction of human partial scans and considers the irregularity of point cloud data structures.
We have improved the network performance in the spatial domain and managed to create better completions using loss function terms and modifications, architectural changes and post processing on the animations. We used elaborate benchmarking and validation tools, to achieve improvements both in quality and quantity.
Project Title:
Students:
Nir Mualem and Oren Shavit
Supervisors:
Amit Bracha and Ron Slosberg
Description:
During a sinus surgery, the surgeon needs to see the
patient’s CT scan.
To make it more convenient, the CT scan will be displayed as an STL 3D file aligned with the
patient’s face using AR with Microsoft HoloLens 2.
The solution must include engineering parts and algorithmic parts for the registration problem as well.
We developed an app using Unity platform, which should solve it and can help the surgeon during the operation.
Code: https://github.com/nir6760/HL2proj
Project Title:
Students:
Supervisors:
Amit Bracha and Noam Rotstein
Description:
While traditional cameras are constrained to two dimensional images, it is possible to utilize several cameras to create systems capable of capturing three-dimensional images. Such systems are called stereo camera systems, and are capable of capturing the depth of every point in the scene, using a stereo matching algorithm and the distance between the cameras. In this project, I created a stereo camera from scratch, which includes wiring the cameras for synchronized hardware triggering, and building an interface capable of streaming and capturing depth images, generating and viewing point clouds, and interactively performing internal and stereo calibration of the camera array.
Project Title:
Weakly Supervised Learning Methods for Analysis of Histological Whole Slide Images
Students:
Supervisors:
Description:
In the field of histopathology, slices of tissue samples are typically stained and examined under a microscope for diagnosis of cancer and other diseases, and for analysis of different properties of the diseases. In recent years, the use whole slide imaging technology for digital scanning of glass slides has rapidly increased, allowing the employment of artificial intelligence (AI) methods to improve the analysis and prediction abilities in computational histopathology. Scanning of a glass slide results in a scan known as a whole slide image (WSI), a gigapixel image of a tissue sample constructed from multiple different magnification resolutions. Performing analysis of WSIs using deep learning algorithms poses multiple challenges, for example due to the large size of the images, and due to the lack of availability of local annotations.
In this project, we implemented and assessed the potential of methods in weakly supervised and self-supervised deep learning to perform analysis and predict clinical outcomes from hematoxylin and eosin (H&E) stained WSIs of patients with breast cancer. We proposed and evaluated ways to deal with the challenges that arise when working with WSIs and applied our methods to a WSI dataset for classification of hormone receptor status information.
Project Title:
Precise Completion of Deformable 3d Shapes Using SIREN
Students:
Supervisors:
Description:
With VR
headset’s popularity growing and its application for VR chats raising interest, the need for a fast, cheap and quality way to transform the
user’s shape and movements in real time to the VR space arises.
In this paper I will propose my idea for a network that can reconstruct a partial scan from a single angle which came from a depth sensor to a full precise 3D shape.
I will discuss the representation of a 3D shape as an SDF and its possible advantages for shape completion, along with SIREN network that uses an easy to calculate Loss function and shows how the use of periodic activation functions such as Sine can improve both the speed and the quality of fitting a point cloud to an SDF.
Finally, I will show how I used SIREN network and SDF representation to implement a new network for the use of precise completion of deformable 3d shapes.
Project Title:
Students:
Supervisors:
Description:
The Robotics Excellence Program in mechanical engineering department are trying to build a robotic arm that picks tin cans (as the picture on the right) from a clutter in a box and organize it for shipment.
Above the box there is a static RGBD LIDAR camera.
Our project was creating the
“vision” part of the robot, meaning each time the robot wants to take the next can from the clutter
– he sends a request to our program that is connected to the static camera that will return the position of the most upper can.
Project Title:
Students:
Yaniv Holder and Ofir Florentz
Supervisors:
Yaron Honen & Boaz Sterenfeld
Description:
A security software which can recognize faces that are in a certain data base in a video broadcast.
Project Title:
AR real-world-object interactions
Students:
Mousa Arraf and Patrisious Haddad
Supervisors:
Honen Yaron & Boaz Sterfeld
Description:
we created an augmented reality game, with real-world-object interactions. two players can connect over the network and play with each other on a playing field that they see through AR. each player has a physical cube that he can move to fend incoming fireballs off.
Project Title:
Students:
Noa Pariente and Barak Biber
Supervisors:
Yaron Honen & Boaz Sterenfeld
Description:
Drum Legends is a virtual reality drummer game.
In this game you drum on a virtual drum set which includes even a bass drum!
To play the bass we constructed a DIY bass pedal which anyone can construct at home using basic materials and a Bluetooth mouse as the actual controller. Everything is affordable and easy to acquire.
In the game you can play your favorite songs in front of your fans and have a lots of fun.
Project Title:
Utilizing Prior Knowledge for Non-Rigid Shape Completion
Students:
Supervisors:
Description:
In recent years, researchers have shown an increased interest in 3D human pose and shape estimation.
Most studies in the field relies solely on completion from partial shape without additional information, resulting a limited models that cannot always reconstruct the partial shape precisely.
The study utilized prior based approach for shape reconstruction of human partial scans that significantly improved the performance of existing methods.
Additionally, in this study we developed and applied new technique for sampling from large datasets resulting solid increase of the performance across all tested learning models.
The sampling methodology presented here has profound implications for future studies of machine-learning models that relies on learning from large datasets.
Finally, we designed new visualization tools to explore the shape and the pose manifold of parametric body models and datasets.
Project Title:
Classification and segmentation of medical imaging
Students:
Nir Shopen and Omer Kawaz
Supervisors:
Description:
An implementation of a DML (Distance metric learning) classifier, based on
"RepMet: Representative-based metric learning for classification and few-shot object detection"
Project Title:
Learning Unique Invariant Signatures of Non-Rigid Point Clouds
Students:
Sari Hleihil and Idan Shienfeld
Supervisors:
Description:
We propose a metric learning framework for the construction of invariant signatures of non-rigid 3D point clouds under the isometry transformations group. We leverage the representational power of convolutional neural networks to compute these signatures and show that in comparison with classical methods, we achieve superior results that allow for higher classification accuracy using the invariant signature, and a lower pose dependency, with the additional advantage of much lower complexity, allowing for the calculation of invariant signatures for larger point clouds with orders of magnitude less time, this is achieved without the use of edge information that is commonly used for such applications. Furthermore, our proposed training scheme allows achieves superior classification accuracy than end-to-end trained classifiers using the same architecture.
Project Title:
Students:
Zahi Cohen, Noy Gini, Adriana Dolgin and Silvan Marti
Supervisors:
Amit Bracha and Ron Slosberg
Description:
The goal of the project is to use the augmented reality capabilities of the HoloLens headset for improving medical surgery process. The main task is to adjust a three-dimensional (pre-prepared) CT model on a patient's lower back, while finding the location and angle at which the model should be allocated, so that it fits exactly to the patient's body and medical procedure.
In this way, it is possible to insert a needle into the right place in the patient's body in a more accurate and simple way than before.
The project extensively uses voice commands, spatial mapping construction, and receiving and processing data from the sensors on the HoloLens glasses, as well as using an ICP algorithm for the purpose of fitting the model onto the patient.
Project Title:
Students:
Rotem Elias and Emmanuel Ferdman
Supervisors:
Description:
Bicycle fitting is a process of adjusting a bicycle for a cyclist. A good bicycle fit is the key to improve cycling. Every cyclist should fit his bicycle to his measurements in order to improve performance, prevent long term injuries and optimize the cycling experience. The bicycle fit creates the connection between the rider’s current physical state and what it is they desire to achieve.
The goal of the project was to create an application that allows to upload a short video of the cyclist riding his bicycle and rank the bicycle configurations. Such ranking should give a good initial indication of the current fitting state. The application is using digital image processing algorithms to detect the rider's body, find the coordinates and calculate the angles. In addition to the height, this information will be used by the software to identify and rate the rider's seating configurations.
Project Title:
Students:
Guy Lecker ,Eduardo Abramoff and Ofek Gutman
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
In short, BlackjackVR is a PC game written in C# and implemented in Unity, our chosen graphic engine. It is using an Oculus Rift headset for the virtual reality aspect of the game and a Leap Motion Controller to elevate the gameplay experience with hand movement recognition.
Project Title:
Students:
Shani Bar-Gera and Shani Bigdary
Supervisors:
Yaron Honen and Boaz Sternfeld
Description:
The goal of the project was to create an application that glorifies a chosen object within a given image. After the glorification, the user would have some sort of ability to “play†with the object by performing various manipulations on the object.
We decided to achieve this goal by creating an Android application that allows users to select an image, or to take one, mark an object in that given image and perform the glorification process. In this case, glorifying means blurring the background and magnifying the object. After the object is glorified the user can perform manipulations on the object such as scaling and rotating. We also decided to add a bonus option - allowing the user to select two objects at once.
Project Title:
Forest Card Wars - A WebGL VR Multiplayer Card Game
Students:
Sagi Taizi and Shachak Gil
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
Our game is a multiplayer VR game built for direct browser play, with no downloads or installs required. We developed an action packed, interactive and competitive two player card game that can be played with any VR headset, by simply opening a browser and typing in the address.
The game draws inspiration from popular card games like Hearthstone, where the main goal is to damage your opponent and deplete his life points before he depletes yours, but with a fun VR twist that involves physical actions the player needs to do.
Project Title:
Aircraft Wing Shape Analysis by On-board Cameras and Deep Learning
Students:
Alexander Portiankin, Ido Plat
Supervisors:
Description:
The main drivers in today's aircraft design are performance improvements, reduction of fuel consumption, and harmful emission. We can straightforwardly achieve these goals with lightweight, large wingspan designs. However, such configurations are inherently more flexible and susceptible to adverse aeroelastic phenomena, including reduced control authority, increased maneuver loads, excessive response to atmospheric turbulence, and flutter instability. The immediate remedy for aeroelastic problems is stiffening the structure.
In this project, we make an essential step - we implement and test a methodology using synthetic, computational data. We simulate a large dataset of 2D wing images and their corresponding deformation parameters and train a neural network, validating it on many unseen examples. We show promising results in preparation for the experimental phase, where this methodology test empirically under lab conditions.
Project Title:
Students:
Supervisors:
Yaron Honen and Alon Zvirin
Description:
Analyzing sports games is a very important task for trainers and players, who wish to improve their skills.
In my project, I present an end-to-end model, that can process and analyze official tennis matches. The model can detect and track the players, the ball and the court, it can also predict the bottom player stroke`s type, calculate all kinds of statistics of the gameplay and create a top view model of the game.
This entire processing can be done on a standard computer with reasonable time.
Project Title:
Students:
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
Climb your way to victory!
VR climbing simulator. realistic feeling of climbing and fear of falling.
unity based game that features 2 main quests:
-Snow Rescue Mission.
-Jack and the Beanstalk.
The game have realistic body to hand mechanics and motion for true climbing
feeling including haptic feedbacks to the hands.
Project Title:
Students:
Efrat Israel and Akiva Zonenfeld
Supervisors:
Description:
Our project is the product of a collaboration between the Geometric Image Processing Lab (GIP) at the Computer Science department and the Educational Neuroimaging Center (ENIC) at the Technion.
The goal of the project is to create a desktop application that analyzes children’s emotions during virtual classes and indicates the children that express negative emotions in real time. In addition, we provide a log with the emotions of each child during the class, including times that face was not recognized, and statistics of the class emotions as a group. The log is used for research in ENIC.
We track the children along a video using face recognition and image processing tools, and output their emotion using a classifier. We are deploying a state-of-the-art model which was developed by the GIP lab for facial expression recognition on children, RonNet.
Although the name of the project indicates that the video conference must be held in Zoom application, our desktop application can be used with any kind of video conference application as well as with video recording.
In order to track each child as an individual we use text recognition tools and match the reappearing name with the relevant adjacent face and emotion.
Project Title:
Augmented Treasure Hunting
Students:
Ilya Freidkin and Michael Cudryavtsev
Supervisors:
Roman Rabinovich and Ibrahim Jubran and Yaron Honen
Description:
This is a game which uses cutting age technologies of AR. Who will find treasure first? The treasure is a virtual object planted in the previously created 3D point cloud of the scene. Using in-door drones couple of players compete to find treasure. In real-time, players can see the live feed from the drone, along with the virtual objects that are hidden in some parts of the map, and rendered into the camera stream.
How to play?
First we need to make a preparation flight - in this flight we will build and save a point cloud map and decide where to locate our "treasure". Once it's done, each player will load our map with the treasure and will start to search for it using drone which controlled by player from a keyboard. This game can be a multiplayer, each player will search for the treasure from his machine, where was installed our game.
Project Title:
Students:
Supervisors:
Guy Koplovitz and Boaz Sternfeld and Yaron Honen
Description:
A ray tracing shader which models cook torrance and black body models.
The user can control the diffusive and specular ratio of each object, as well as the temprature of the black body objects.
Project Title:
High Perceptual Quality Image Denoising with a Posterior Sampling CGAN
Students:
Guy Ohayon and Theo Adrai
Supervisors:
Gregory Vaksman and Michael Elad
Description:
The vast work in Deep Learning (DL) has led to a leap in image denoising research. Most DL solutions for this task have chosen to put their efforts on the denoiser's architecture while maximizing distortion performance. However, distortion driven solutions lead to blurry results with sub-optimal perceptual quality, especially in immoderate noise levels. In this paper we propose a different perspective, aiming to produce sharp and visually pleasing denoised images that are still faithful to their clean sources. Formally, our goal is to achieve high perceptual quality with acceptable distortion. This is attained by a stochastic denoiser that samples from the posterior distribution, trained as a generator in the framework of conditional generative adversarial networks (CGAN). Contrary to distortion-based regularization terms that conflict with perceptual quality, we introduce to the CGAN objective a theoretically founded penalty term that does not force a distortion requirement on individual samples, but rather on their mean. We showcase our proposed method with a novel denoiser architecture that achieves the reformed denoising goal and produces vivid and diverse outcomes in immoderate noise levels.
Project Title:
Reinforcement Learning Approach for Formula Driverless Car
Students:
Supervisors:
Yaron Honen and Roman Rabinovitch
Description:
Formula Student Driverless is a competition between universities all over the world to build an autonomous race car that will drive successfully in a racetrack, with events and competitions happening all throughout the year in many cities around the world.
In this project I implemented Reinforcement Learning algorithm called SAC (Soft Actor Critic) on the simulated Technion autonomous Formula car, with simulation done in Unreal Engine 4 and AirSim plugin, and input being only an image from a camera mounted on the car, while the car has no knowledge or assumptions on the shape of the track.
I used a VAE (Variational Auto-Encoder) to reduce image size and thus reducing the observation space, and the output of the algorithm is a steering angle, keeping a constant speed of 7.5 m/s (27 km/h).
My work in the GIP Lab in the Technion - Israel Institute of Technology, strives to contribute to the amazing Formula Student Technion Driverless team in their efforts to compete in future events.
My code in GitHub: https://github.com/eliork/Reinforcement-Learning-on-Autonomous-Race-Car
Project Title:
Students:
Supervisors:
Alon Zvirin and Yaron Honen
Description:
Chest motion and respiratory volume abnormalities or sporadic breathing rate are often associated with thorax diseases hidden under the human eye radar. Those can range from light and passing conditions to fatal and lifestyle affecting illnesses, which in many cases stay unnoticeable or falsely diagnosed.
Current medical methods of chest motion abnormalities detection often rely on human medic eye observation which might be less accurate and comprehensive as computer vision, and often lead to mistreatment of patients who are unable to be physically present at the examination location due to various reasons: handicap, privacy or mostly - availability.
Our work in the GIP Lab in the Technion Institute strives to provide an innovative, precise and accessible-to-all tool to detect chest motion and breathing abnormalities as seeked by medical doctors and provide the findings as fast as possible, using a 3D camera.
Project Title:
Corn leaf segmentor - Phnomics
Students:
Adva Cohen and Shani Idgar
Supervisors:
Yaron Honen and Alon Zvirin
Description:
Analysis of maize leaves is a widespread issue, important for assessing plant growth. In our project our goals were to improve segmentation of maize leaves and to classify maize plants into two categories, untreated and fungi-infected, using our segmentation to create the dataset. Our methods to improve segmentation included a two-step inference process and improving the training by creating synthetic images. Our methods for classification included creating a Cifar-10 based CNN architecture model, trained from scratch. We demonstrate that creating a larger dataset using data augmentation and training the networks from scratch improves both segmentation and classification.
Project Title:
Semantic Segmentation of Cloud Images Using Weakly-Labelled data
Students:
Roi Tzur-Hilleli and Aviv Caspi
Supervisors:
Reut Yehonatan and Yaron Honen
Description:
Semantic segmentation is the task of labeling each pixel in an image with the class of which the pixel is a part of. In order to train a deep convolutional network that will accomplish this task, it requires a lot of hand-drawn segmantation maps and therefore a lot of resources.
We wanted to check whether it is possible to save resources on the manual labeling of cloud segmentation maps and still achieve competitive results.
In order to find out, we trained a deep convolutional network for semantic segmentation of cloud images on both fully-labeled data (full segmentation maps) and weakly-labeled data (scribbles) and compared the results achieved by both methods.
Project Title:
Students:
Supervisors:
Description:
YOGA MASTER is a platform for yoga-practicing anywhere you want. With the Jetson Nano, you can take the yoga teacher to the park, to the beach, or just stay at home. Although the teacher is not near you can still get the feedback you need to improve your poses.
Project Title:
Students:
Shlomi Ovadia and Hila Akerman
Supervisors:
Description:
Super Mario Go is a mobile game, harvesting Unity's AR Foundation plane detection capabilities to create an 'On Surface' playground.
providing two different game modes:
1. classic mode - the player can play a classic pre-designed course and place it on the 'physical' environment.
2. survival mode - the player can walk around freely in the 'physical' environment, fight and avoid enemies and trying to get a high score.
Project Title:
Students:
Zohar Rimon and Adi Arbel
Supervisors:
Description:
Motion Transfer, the task of reenacting the image of a person according to the movement of another, is an active research field in computer vision. Recent methods achieve realistic looking results in controlled scenarios. Yet it is difficult to obtain similar results for complex, crowded, in-the-wild scenes while integrating the reenacted person into the target scene. In this work, a novel workflow is proposed to tackle this challenging scenario, which we name Scene-Aware Motion Transfer (SMT). Our workflow harnesses a set of models, each attaining state-of-the-art results in its respective field, and is divided into two major stages. First, a novel person tracking pipeline is used to separate each unique identity from the crowd. Then the tracking results are utilized for a targeted single-person motion transfer, resulting in a fully automatic workflow that can handle complex videos. An extensive evaluation is presented to show the quality and robustness of the results in different scenarios.
Project Title:
Students:
Michal Guttmann and Amit Shuster
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
We developed a VR game that uses a convolutional neural network (CNN) to identify commands from the user, drawn using the controller. The game was inspired by Harry Potter and the Philosopher’s Stone (2001).
The main goal of our project was to involve a Deep Learning mechanism in the VR development environment. After a lot of research, we decided to build a CNN by ourselves and use it in our game.
Project Title:
Efraim, the Autonomous Farmer
Students:
Annael Abehssera & Yonatan Gershon
Supervisors:
Description:
Efraim is a 3D printed robotic arm that uses an image processing real-time segmentation algorithm named Yolact, an Intel RealSense 3D camera and an Arduino board to autonomously identify & locate an object and pick it using its robotic arm.
Project Title:
Realtime Breathing using realsense
Students:
Nili Stein and Maayan Ehrenberg
Supervisors:
Alon Zvirin and Yaron Honen
Description:
Chest motion abnormalities and sporadic breathing rate are often associated with thorax diseases hidden under the human eye radar. Those can range from light and passing conditions to fatal and lifestyle affecting illnesses, which in many cases stay unnoticeable or falsely diagnosed.
Current medical methods of chest motion abnormalities detection often rely on human medic eye observation which might be less accurate and comprehensive as computer vision, and also often lead to mistreatment of patients who are unable to be physically present at the examination location due to various reasons: handicap, privacy or mostly - availability.
Our work in the GIP Lab in the Technion Institute strives to provide an innovative, precise and accessible-to-all tool to detect chest motion and breathing abnormalities as seeked by medical doctors and provide the findings as fast as possible, using a 3D camera.
Project Title:
Abiotic Stress Detection In Banana Plants
Students:
Supervisors:
Alon Zvirin and Yaron Honen
Description:
The banana plants are very important and are a big part of the nutrition of many parts of the world.
More than 100 billion bananas are eaten every year in the world, making them the most popular agricultural product.
4 different qualities of treatment were applied and documented by a picture in a sequence of 17 days.
In the following project, we will try to distinguish between 4 kinds of Banana treatments by
observing the pictures only.
We will use augmentations and introduce a novel specific data augmentation that is targeted for
our dataset.
We will try to give our classifier an expert's capability since the differences are not notable and
might not be observable to the common viewer.
Project Title:
Low Complexity Data-Efficient Generation in Disentangled Setting
Students:
Shavit Borisov and Jacob Sela
Supervisors:
Description:
Recent advancements in the fields of generation are promising better results, with more control over generation. Unfortunately, these results are achieved with massive data sets used to train highly complex models by the leading experts of machine learning, making them inaccessible to "the average Joe". In this paper, we propose that the disentangled latent spaces created as a by-product of these tools can be repurposed for generation with a specific factor of variation in mind, using simple tools and little data. We demonstrate these claims by generating aging videos using NVidia StyleGAN’s latent space from a single source image.
Project Title:
Students:
Supervisors:
Ibrahim Jubran and RomanRabinovich and Yaron Honen
Description:
This is an Augmented Reality first-person shooter game for Android.
You play as a drone armored with laser guns.
You need to pass through the course of virtual obstacles,
and defeat the enemy drone while striving to get the highest score in the process.
You can create your own game course by configuring images and spreading them across your house. These images will be used to render virtual objects for you to interact with them in the game.
Project Title:
Augmented Reality Laser2Target
Students:
Supervisors:
Description:
In this project, we used Unity combined with different packages (Photon, Vuforia) to create an easy solution for people who want to navigate towards a human-destination.
We provide a simple and clean user-interface that includes an arrow+distance towards the target, and an augmented-reality overlay for faster recognition of your target.
Project Title:
Augmented Reality Indoor Navigation
Students:
Arkadi Gurevich Shahar Zivan
Supervisors:
Description:
Navigation inside a large indoor facility such as an office building or shopping mall is a problem that isn't solved by any modern day navigation solutions, due to the imprecise results of GPS based coordinates. In this project we offer a solution based on augmented reality and image recognition.
Our solution is based on using image recognition to find landmarks in the environment such as posters or billboards and learning the user's position relative to them, thus receiving a precise position to base our navigation on.
Our results show this method can give very precise results and allow navigation in narrow corridors and small offices with no mistake, as long as landmark is always inside.
We conclude that this is a viable solution to this problem, although some position-correcting method is required when no landmark can be seen.
Project Title:
Students:
Roni Ben Dom and Ophir Havivi
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
In this project we created a matrix movie like effect over a pre constructed scene.
Starting from a normal room and by a push of a button the scene changes: it becomes matrix world
Project Title:
Students:
Snir Homsky and Iliya Rubinchik
Supervisors:
Alon Zvirin and Yaron Honen
Description:
Mask R-CNN is a conceptually simple, flexible, and general framework for object instance segmentation which efficiently detects objects in an image while simultaneously generating a high-quality segmentation mask for each instance. This Project tackles the task of corn plant segmentation given a partially annotated small dataset using Mask R-CNN. We present a workflow for producing a large dataset from a small given dataset, which includes image augmentation, generation of artificial plant images, and generation of artificial images simulating real greenhouse scenes. Finally, we present results on leaf segmentation as well as whole plant segmentation, and discuss these results.
Project Title:
Students:
Tal Leibovitz, Aviel Simchi
Supervisors:
Yaron Honen and Boris Van-Sosin
Description:
As the development and usage of VR and AR technologies increased in the last years there isn’t yet a
cheap and precise solution for finger-tracking problem.
This follow-up project designed to continue the work we have done as members of the B-Glove
hardware and software teams. In our main project, we looked for creating a solution for the finger-
tracking problem using IMUs (Inertial Measurement Unit) located on the fingers and the back of the
hand.
With the help of IMUs, we were able to determine the orientation of each finger and the entire
hand. These orientations are sent to the computer which displays a 3D hand in Unity scene.
Project Title:
Children facial expressions detection with EEG from video
Students:
Dana Goghberg, Moran Hait and Sapphire Elimelech
Supervisors:
Gary Mataev and Ron Kimmel and Tzipi Horowitz Kraus and Michal Zivan
Description:
Our project is the product of a collaboration between the Geometric Image Processing Lab (GIP) at the Computer Science department and the Educational Neuroimaging Center (ENIC) at the Technion. The goal of the project is to create an emotion recognition system based on facial expression recognition (FER) algorithm, which will be used by ENIC lab for analyzing children’s responses to various tasks, while being monitored by EEG/fMRI.
We track the children along a video using face recognition and image processing tools, and output their emotion using a classifier.
We used convolutional neural network to implement a new facial expression recognition system which is specifically designed for children, Although well studied on adults, only a few facial expression recognition studies have been done involving children, and consequently only a handful of small relevant datasets exist. This difficulty was added to the fact that some children’s emotions have similar representation, and are therefore difficult to differentiate.
We tested our chosen model using cross-validation and a real-time emotion detection from a video recorder. The results exceeded our expected goals.
We introduce a novel machine-learning system that detects facial expressions in children with 91% accuracy.
We are not aware of approaches yielding similar results in the literature.
Project Title:
CHANGE DETECTION IN REMOTE SENSING SATALITE IMAGES USING DEEP LEARNING
Students:
Ayelet Alon, Inbal Tziperman Lotan and Tal Amir
Supervisors:
Description:
Change detection in remote sensing images is an important part of many applications such as tracking urban changes for military purposes, tracking deforestation for climate change research, agricultural monitoring and more. We pose the change detection task as to identify differences between two photos, taken at the same geographical location at two different times. The detected changes should be a result of appearance of new or disappearance of existent objects in a scene. The images may contain other differences between them that are a result of many factors such as seasonal changes (snow, trees with/ without leaves), changes in brightness, shifted images due to slight changes in the capture angle and many more. Tracking changes manually is a tedious task and can be highly time consuming. In addition, since the task is extremely tedious, doing it manually often results in human errors. The automation of this task is therefore highly desirable. In this project, we present an automation of the change detection task using both a fully and weakly supervised semantic segmentation based neural network. We compared the performance of our model to a previously published GAN based model, and showed that though our model is easier to train it achieves similar results.
Project Title:
Students:
Supervisors:
Description:
Procedural generation is a method of creating data algorithmically instead of manually.
Procedural map generation is a derivative of this method for creating unending worlds.
Imagine yourself as a character in a video game exploring and traveling through a vast virtual world.. when suddenly, you reach the end of this world, a cliff, if you take one more step you will fall to infinity never reaching the ground.
Procedural Map generation comes to solve this issue and allow you to explore the world without it ever ending around you.
My project allows the player to explore a building which is built in a procedural manner around the player online as it moves. The building is built from basic building blocks that were built in a specialized program. These basic blocks are connected according to a set of rules which was written beforehand.
Project Title:
Students:
Supervisors:
Description:
This paper presents the development and details for use of a people’s counting system, able to count the people entering and exiting a specific zone from a live stream video feed obtained from IP cameras in real-time.
This goal is achieved by the combination of two components: a detector component using a Convolutional Neural Net (YOLO) detecting people on the frame, and a tracking component utilizing a tracking algorithm (CSRT) which updates those people positions on the next frames.
This system was built for the Technion’s libraries to monitor the amount of people in the libraries at any given time but is highly configurable and can fit different types of building and entrances.
Project Title:
Augmented reality in road navigation
Students:
Supervisors:
Ron Slosberg and Gill Shamai
Description:
We have created an AR road navigation guidance system
to assist during driving. Our system uses deep learning to
by pass the need for GPS signal and translates Google
maps instructions into a visual arrow. This augmented
arrow points towards the designated path on the screen of
a smartphone’s camera.
Project Title:
Physical Cloud Generation Simulation
Students:
Supervisors:
Revital Konch and Yaron Honen
Description:
We generate real-time volumetric clouds in Unity using ray marching.
The simulation has many parameters that influence the look of the final result and can be controlled by the user.
The final simulation can be imported as an easy to use Unity package.
Project Title:
Mindfulness in Virtual Reality
Students:
Ola Awisat and Hala Awisat
Supervisors:
Michal Zivan and Yaron Honen and Boaz Sterenfeld
Description:
Application that allows the user to practice mindfulness in a virtual reality environment. It uses muse headband to measure real time EEG data of the user, and triggers changes in the environment based on this data, creating a neurofeedback loop. the main purpose of the application is to help the user become more mindful on the long term.
Project Title:
Neuro feedback based Virtual reality for practicing mindfulness principles
Supervisors:
Description:
In this project we would like to expend the previous project by adding more channels of physiological input from the user together with more sensory feedback to the user. In addition new VR scenes will be designed.
The project’s main objectives are -
1. Getting to know the current system.
2. Integrate new physiological inputs of Heart rate and breathing (which are additional outputs of the muse 2) to the system software.
3. integrate a tactile sensory feedback to the user.
4. Create new VR scenes.
Project Title:
Students:
Sagi Barazani and Adi Reznik
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
In the project, we demonstrate proof of concept for replacing daily physical actions with actions that are done via gestures in augmented reality. The demo was done on LIFX - smart lamp and Spotify - music player.
With the help of these capabilities, children, people with disabilities, and adults can easily perform activities that in the past required physical effort. The next step will be to connect the furthermore smart devices in our physical world so that we can control them through the virtual world.
Project Title:
RespiTrack - Respiratory patterns tracking Android App
Students:
Guy Berger and May Schwartz
Supervisors:
Alon Zvirin and Yaron Honen
Description:
A
person’s breathing pattern abnormalities are serving a major role in detecting many human diseases like asthma, acute respiratory failure, heart failure and more.
Yet the detection and analysis of such pattern abnormalities
isn’t yet as advanced as other advancements in medicine and in most of the cases is subject to human analysis by a human and thus involves 2 major drawbacks which we aim to solve:
Physical examination - the patient must arrive at a clinic and be examined physically by a medical staff member.
Human error - The diagnosis is made by a medical staff member and thus prone to human mistakes.
The work that is being done by the GIP (Geometric Image Processing) Lab at Technion university in Haifa is the development an innovative automated system and method to detect normal and abnormal breathing patterns by performing analysis of breathing pattern pathologies and abnormalities in chest movement.
In this project we develop an Android App that follows this methodology and aims to solve the drawbacks described above.
Project Title:
Students:
Almog Brand, Dani Ginsberg and Lior Wandel
Supervisors:
Yaron Honen and Boaz Sternfeld and Boris Van Sosin
Description:
Augmented reality shadows are un-realistic compared to the shadows of “real
world†objects created by the “real world†light module.
In Order to build a light model that we can deploy to the AR system, we first needed a way to measure the room’s ligh
Project Title:
Facial Expression Generation using GANs
Students:
Supervisors:
Yaron Honen and Gary Mataev
Description:
The main goal is to generate synthetic data for projects, in the machine learning field, that deals with face emotions classification.
To classify images with multiple class labels using only a small number of labeled examples is a difficult task. Especially when the label (class) distribution is imbalanced. In face emotion classification we have imbalanced label distribution because some classes of emotions are relatively rare comparing to others. For example, disgust emotion is more rare than happy or sad.
In this work, we propose a data augmentation method using generative adversarial networks (GAN). It can complement and complete the data manifold and find better margins between neighboring classes. Specifically, for this task, we are using classifiers based on Convolutional Neural Networks (CNN) and a variation of cycle-consistent adversarial networks such as CycleGAN, Improved CycleGAN, and The Wasserstein CycleGAN. The CycleGAN is a direct implementation of Emotion Classification with Data Augmentation Using Generative Adversarial Networks paper.
In order to improve the results and avoid problems that we faced we employ different variations of CycleGANs. We show, that our empirical results can obtain a ~5% increase in the classification accuracy, after employing the GAN-based data augmentation techniques.
Project Title:
Students:
NitzanWinkler and Bar Neuman
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
We developed a workout VR game. The game includes a tutorial stage, meant to teach the game. A training option with no timing for practice option. And finally a stage where you can play for 1 minute trying to achieve the highest score you can
Project Title:
Bringing meditation to virtual reality
Students:
Ola Awisat and Hala Awisat
Supervisors:
Michal Zivan and Yaron Honen and Boaz Sterenfeld
Description:
This project is a part of the collaboration between Geometric Image Processing Lab (GIP) and the Center for Graphics and Geometric Computing (CGGC) at the Faculty of Computer Science at the Technion, and the Educational Neuroimaging Center (ENIC) at the Faculty of Education in Science and Technology at the Technion. The project aims to create an electroencephalography (EEG) based brain-computer-interface (BCI) that aims to provide an environment to practice mindfulness in virtual reality. The system we developed employs a virtual reality environment consisting of a realistic sea environment. The system acquires EEG data in real time and determines the mental state of the user. It then triggers positive feedback in order to reinforce the mental state of relaxation, and negative feedback that brings the attention back on relaxing. EEG data acquisition was done using Muse – an EEG device that measures the brain waves in real time and transmits them on Bluetooth. Muse uses signal processing methods to transmit spectral power in alpha, beta, gamma, delta and theta frequency ranges along with the raw EEG data.
Project Title:
Students:
Orel Haim, Shai Yaakovi and Lital Wexler
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
As part of our B.Sc. in Computers Science, we developed a VR game that based on the nostalgic Dancing Carpet game. Our application uses 2 trackers that are attached to the user ankles and simulates a dancing carpet on the floor, the user is playing and chosing his preferences in a specialy design menu only by moving his feets.
For an extraordinary user experience the game contains couple of diffrent enviroments, difficulty levels and songs.
Project Title:
Students:
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
an AR application which translates Sign Language, it captures the sign from the Gloves, translates and projects it live in Microsoft HoloLens headset
Project website : https://aseelsakas.wixsite.com/website
Project Title:
Students:
Naama Glicksberg, Guy Lazar
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
We are Gai and Naama, students in the last semester in the computer science faculty in the Technion.
When we heard about the project in the lab, which includes building a virtual reality game, we were very excited and decided to choose it.
When we thought of an idea for a game, the idea of Tetris came up - a simple and fun game that has been familiar to everyone for years.
We thought that turning the game into 3D, so that the game cubes are flying around the room in front of the player could be very interesting and fun!
Project Title:
Students:
Batel Carmona, Mor Eliyahu and Lev Pechersky
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
Pacman of thrones is a VR remake of a classic Pacman game, using virtual reality and game of throne motifs.
In our game the user plays as Pacman, who controls the game and movement like the classic Pacman game, but of course we added a few features of our own, now the user can jump, kill the ghosts, and move freely around the maze.
We created a remake of the famous game Pacman in 3D environment for HTC VIVE VR. While developing the game, we tried to add multiple game mechanics in order to give the user a full experience of a true reality.
We managed to create this experience by adding animations of 3D objects, adding many player’s gestures which are closed to the real reality, like: movement, jump and attack.
Project Title:
Real world physics VR experience
Students:
Gil Beresi, Gil Blinstein
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
Our main goal in this project was to simulate real life physics experience, using HTC VIVE headset and trackers.
We wanted the explore the virtual world by doing various actions using hand movements only, i.e no triggers involved.
We have successfully met our goals, by developing reliable and efficient throwing algorithm that immitates a real life throw in the virtual reality.
Project Title:
Students:
Or Eli Pilosof, Rafi Cohen, Yael Tsafrir
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
As part of our BSc in Computer Science at the Technion, we heard about the AR experience with HoloLens and wished to try it ourselves.
When we began exploring the possibilities of the HoloLens×”, we recalled one of our favorite childhood TV series: Yu-Gi-Oh!
In Yu-Gi-Oh!, the characters duel each other with monsters, and the monsters are summoned using special cards.
We created our own version of the Yu-Gi-Oh! game, where players use cards from the original card game to summon monsters in AR, and duel against an computerized opponent.
To make the AR experience more alive and real, we added shadows to the monsters in the game, and gave the player control over the direction of the light in the scene.
Project Title:
Students:
Gal Shalom , Ariel Iny , Alex Bondar
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
As students who learn B.Sc. in Computers Science, we aspire to use computers and programs to make complex tasks
simpler and easier to do.
We decided to develop an augmented reality application which will teach people to play a piano. We used Hololens'
AR properties as our AR platform of development.
Our application uses the Hololens' webcam to recognize the piano keyboard. We used image processing to be able
to recognize the piano and to detect the press of a key. While learning to play a song, the application will color the
next key to press and will indicate with color change if the correct key was pressed.
We hope this application will be an example of the possibilities the AR world has to offer.
Project Title:
Leap Motion Matrix - Multiple Devices
Students:
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
While browsing Leap Motion desktop applications we have noticed that they all have a limited field of view. When we saw the new multiple device beta API release we have decided to try and upgrade it.
As part of the project we worked to synchronize the devices, we built a wooden matrix for the cameras, built a 3D model as a case for the cameras and dealt with numerous engineering issues with the matrix, the cameras, and with some beta-version related programming issues.
And so, we have succeeded in upgrading the field of view, linear to the number of camera devices.
https://danielmines111.wixsite.com/matrix
Project Title:
Facial Expression Classification
Students:
Supervisors:
Elad Richardson and Yaron Honen
Description:
In this project we implement a facial expression classifier. Our facial expression classifier matches the user's facial expression to the corresponding emoji. This project, called Keymoji++ is meant to expand upon the original Keymoji project where we created an android keyboard that uses the phone's front facing camera to detect the user's emotion to be either happy, sad, surprised or angry by their facial expression and recommend the corresponding emojis above the keyboard. In this project we expanded upon the idea and added support for many different kinds of emojis.
Project Title:
Students:
Lorella Matathia, Tal Albo, Dan Simah
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
As part of our B. Sc. in Computers Science, we were curious about the Virtual Reality world. As we tried for the first time the Oculus technology, we immediately thought about the possibilities for a VR RPG game, and the amazing games that could possibly come.
We created an "Escape Room" - like game with the emphasis of creating generic building blocks for future VR applications. We truly hope that this game will inspire the next generation of RPG gaming and help with the development of VR application.
Furthermore, we would like to thank to the GIP & CGGC staff, and especially to Boaz Sterenfeld and Yaron Honen.
Project Title:
Cucumber Detection and Segmentation
Students:
Asher Yartsev Or Shemesh Simon lousky
Supervisors:
Description:
This Project tackles the task of segmentation and indexation on cucumber plant parts. The approach is to implement an end to end workflow to partially annotated and relatively small datasets.
Mask RCNN paper from facebook research has proven very efficient in this task even on small datasets. Therefore the focus is maximizing the effectiveness of a problematic dataset on different levels. This reports discusses partially annotated datasets, and inconsistent annotation styles.
The conclusion tends towards a trade-off between investing in the dataset quality and using augmentation acrobatics to quickly use the existing data. There is no doubt that consistent fully annotated real pictures will get the most out of the network.
Project Title:
Students:
Aviel Simchi Amit solomon
Supervisors:
Boris van Sosin and Yaron Honen and Boaz Sternfeld
Description:
As the development and usage of VR and AR technologies increased in the last years there
isn’t yet a cheap and precise solution for the finger-tracking problem.
In our project, we looked for creating a solution for this problem using IMUs (Inertial Measurement Unit) located on the fingertips and the back of the hand. With each IMU we were able to create a 3-dimensional axis system and thus determine the relative direction of the hand and each finger to the ground and to each other.
As the HW team, we provided a glove with the abilities described above and a c# adapter that can be used in unity environment.
Project Title:
Students:
Michal Guttmann and Amit Shuster
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
We developed a VR game which was inspired by several games including BBTAN by 111% - Mobile game, Hurl VR - VR game and Let's Bounce - Arcade game.
Our application simulates several environments with diverse difficulty levels. In each level, you have colored cubes organized in a different configuration. The player's main goal is to throw balls at the cubes and hit all the colored cubes on the level, by hitting a cube, the cube is turned black. The more balls the user used the less score the user will get, namely, the user strives to minimize the number of used balls.
Score diverse from 1 to 3 stars while 3 stars are the best. The player can enjoy both the game and the VR experience using VR goggles, trackers and gloves. The application gives the player an extraordinary experience that enables him to use several senses in order to get a high score.
Project Title:
Facial Reconstruction from Video
Students:
Maxim Aslyansky Volodymyr Polosukhin
Supervisors:
Ron Slossberg and Elad Richardson
Description:
The goal of the project is to develop an algorithm for face reconstruction from monocular RGB video. Recently it has been shown that neural networks can achieve accurate results for single-image facial reconstruction, outside the scope of limited linear models. In this project we will explore the world of 3D Face Reconstruction from video by using the following approaches:
1. Extending the 2.5D output of current single-image methods with state-of-the-art geometric algorithms for dynamic 3D reconstruction to incrementally build a complete high-quality facial model.
2. Experimenting with novel deep learning architectures to directly regress 3D models from images and also using recurrent neural networks to make use of temporal information.
Project Title:
Students:
Nir Chachamovitz and Ilan Doron
Supervisors:
Description:
In this project we introduce an implementation of an open-source system for automatically recognizing and translating text in pictures, posters, road signs etc. Unlike in existing projects, we focused on preserving the original text style in order to create visually-pleasing results. In addition, effort was made to maintain and preserve the original background of text containing regions within the images.
The project explores:
Deep models for text detection and recognition: We used an already-trained text detection and recognition networks, and a few more deep models in several steps of the pipeline.
Style transfer and image blending techniques: We used perspective transformation technique and other image blending tools.
Project Title:
Students:
Supervisors:
Ron Slossberg and Roman Rabinovitz
Description:
Autonomous drone that pops balloons
Project Title:
Efficient Restoration By Compression
Students:
Yuval Shildan, Nevo Agmon, Danny Priymak
Supervisors:
Description:
When dealing with signal compression, most compression algorithms optimize the reconstructed output with respect to the acquired input signal.
A more general approach would be optimizing the compression algorithm using the signal prior to the acquisition phase as input, such that the final decompressed output is optimal.
Ideally, this approach could utilize knowledge about specific acquisition devices to further optimize the compression. For example, given a known degradation model of a specific camera as an acquisition device (which can be easily modeled by the manufacturer), the presented approach could yield a highly optimized compressed result, which can be either used to transmit a signal of higher quality over the same infrastructure, or alternatively, deliver the same quality using less resources.
Dar et al. [1] proposed an algorithm for joint restoration and compression of images, on which we rely in this work. This algorithm was implemented in MATLAB, which, due to MATLAB's overhead, opened up the possibility of significant efficiency improvements. For this reason, we chose to focus our efforts into optimizing the implementation runtime demands, while considering software engineering and object oriented design as top priorities.
[1] Y. Dar, M. Elad, and A. M. Bruckstein, “Restoration by Compression,†IEEE Transactions on Signal Processing, vol. 66, no. 22, pp. 5833–5847, 2018.
Project Title:
Students:
Tarik Sirhan, Sabah Saloty
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
VR Keyboard application for typing texts, using ManusVR gloves that provide a realistic interaction to the user.
The keyboard contains English Alphabet, Numbers, Emojis and special keys such as Enter, Caps Lock, Backspace, etc.
The application may be added to several VR games in order to be used as a chat interface between players, or may be used for getting input from the user.
Project Title:
Students:
Adam Elgressy & Dmitry Vlasenko
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
Our main goal was to simulate a first person Spiderman experience, using the HTC Vive Pro kit. We also put a goal for our self, to Learn about the new, and evolving world of VR, and gain hands-on experience, with the newest technology, alongside getting to know the Unity environment.
We have successfully made a VR application, which simulates a first person experience, of one of the most famous superheroes of all time, Spiderman. We have made it possible to walk, climb, and swing from a web, being shot from the in-game hand, all while being able to stay stationary, at a one place, but fill like you can explore the whole world.
We have simulated a physics system, which made the experience feel authentic, from the free fall acceleration, to friction with different surfaces, and many more.
We would suggest for future work to be done to add different scenarios, a leaderboard, tutorial room.
Project Title:
Students:
Netanel Lev & Eitan Kosman
Supervisors:
Description:
We developed an end-to-end autonomous car system that operates in a driving simulator.
Project Title:
Students:
Supervisors:
Description:
Humans have always been communicating nonverbally using their body movements
and facial expressions. Those gestures tend to deliver the speaker's message and
feelings much more effectively than the words spoken. Nowadays, with instantmessaging
becoming the common way of communication, those traditional gestures
are represented by tiny faces called emoji. Each sent emoji represents the sender's
emotion or state of mind. Those emojis are manually selected by the sender. We
offer a new advanced automated way of using emojis. By supplying a fully
functioning Android keyboard that captures a shot of the sender's face at a given
time, analyzes his expression and state of mind, we can offer an emoji represents his
feeling.
Project Title:
Students:
Denys Svyshchov, Alexander Gurevich
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
Rubik’s Cube Trainer is VR application the main purpose of which is step by step learning of solving
Rubik’s Cube and opportunity to solve cube on time and compete with friends who is the best Cube solver. The game uses VR gloves allowing the player realistic interaction and better understanding of solving process.
Project Title:
Students:
Eden Ben Oz & Ori Shem-Tov
Supervisors:
Yaron Honen and Boaz Sterenfeld
Description:
We developed a VR environment for drawing in 3D.
The player can draw with both hands using controllers, and can edit their drawings by resizing, moving or deleting them.
The player gets a unique drawing experience, such that they can draw
“on air”, walk through their drawings, catch and move them, and so on.
Project Title:
Students:
Yonatan Sherwin,Roee Peleg
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
Kinetic Painting is a unique VR application for drawing using hand gestures.
With no physical controllers at all, you’ll be able to choose between two main painting methods.
The user will be able to start the painting, erase it, change the level, change the colors of the paint, and even create some special effects.
Using the leap motion controller, you can control and activate the game with various of hand gestures.
Project Title:
Students:
Or Gitli,Mirit Alush, Tal Pilo
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
Advanced whack a mole is a VR game based on the common arcade game “Whack a mole”. The game includes two levels: The first one is a classic version of the common game “Whack A Mole”, well known from the 90’s arcade rooms. The impact is with a hammer that appears when the player make a fist with his right hand. The second one is an advanced version of “Whack A Mole” includes horizontal view of about 120 degrees. The player stand in front of seven pillars, and ghosts come throw from them. The goal is to hit the ghosts with his fists.
Project Title:
Students:
Snir Zango ,Rotem Ohana,Or Harchol
Supervisors:
Yaron Honen and Boaz Sterfeld
Description:
A VR game that is based on
the nostalgic Mega-Son
game "Duck Hunt"
The player's main goal is to
kill as many ducks as
possible and get points, the
speed of hitting the ducks
is also a factor
We added another
environment to the game
(that wasn't in the original
game), a zombie mode,
and adjusted both to the
VR experience
Project Title:
OLM - object location memory
Students:
Bar Neuman, Nitzan Winkler, Jonathan Weizman
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
VR Platform for a research following a demand made by a researcher in psychology from Emek Izrael Institution.
The experiment focused on the subject's ability to memorize the objects in the environment and identify the changes in their positions.
The environment simulates a Savanna with bushes meant to be memorized and foxes as a distraction.
Project Title:
Students:
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
The matrix effect was introduced in The Matrix film that was released in 1999.
It was one of many novel new visual effects that were introduced in this film.
In this project, I made the Matrix rain code effect by creating a Shader in Unity.
Project Title:
Tomato Classification using DL
Students:
Supervisors:
Description:
Diagnosis of crop phenotypes is a widespread problem that affects the life of millions around the world. In our project, our goal was to create a system for classifying different parts of images in order to achieve classification and segmentation of significant plant parts. Our methods included applying neural networks. We've tested a few types of networks, mainly classifier and encoder-decoder. Using the classifier, we achieved very high accuracy (97%- 98%) when classifying parts of the image. Later, using the encoder-decoder architecture we shortened the required time for segmenting an image, from minutes to a few seconds.
Project Title:
Students:
Daniel Weisberg Mitelman, Lorella Matathia
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
360 Video Editing is a unique VR tool that gives you an option to edit a video, without losing the special experience. With the controllers, you can interact with the editor. You are able to delete frames, show specific frames and export the edited video.
The main functions that the editor provides:
• Displaying a video on sphere
• Dividing the video into sub-frames
• Displaying sub-frames
• Splitting video
• Showing effects
• Exporting final video
Project Title:
Students:
Sapir Mordoch, Bar Uliel, Moran Nisan
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
Beat Saber is a unique VR game where your goal is to slash the beats (represented by small cubes) as they are coming at you.
The game includes several songs, each with different difficult. The player uses VR motion controllers to wield a pair of light-sabers to slash the blocks. Each block is colored red or blue to indicate whether the red or blue saber should be used to slash it. When a block is slashed by the suitable saber, the block is cut and the player get a point, otherwise if the player slashes the block with a saber with the opposite saber the score is reset.
Project Title:
Students:
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
A multiplayer space sim game.
Server-Client communication implemented from scratch on top of Unity's low-level networking API (Transport Layer API).
Movement is smoothed with linear interpolation.
Project Title:
Students:
Shlomit Sibony, Ran Mansoor, Yuval Shildan
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
Create your art via our 3D VR paint.
Use the latest technologies including the Manus-VR gloves (wearables) and the HTC Vive.
We offer an intuitive and powerful platform which you can paint with both hands, control the painting transform and choose the colors you wish to paint with.
Project Title:
Quad-copter GIP Vision Control Farmework
Students:
Nathan Sala and Ofir Zelig
Supervisors:
David Dovrat and Ohad Menashe and Roman Rabinovich
Description:
Drones can be used to various tasks and replace humans in many aspects of our lives. Our goal was to create a system that will enable to control a drone using computer vision that will enable it to fly autonomously. Eventually, we created a framework that wraps all the parts that control the flight functionality of the drone in a way that future projects could use it could focus only on computer vision.
Project Title:
Students:
Netanel Lev & Dolev Ben Ami
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
Design, explore and share floor plans in VR.
Project Title:
Child facial expression detection
Students:
Eden Benhamou and Deborah Wolhandler
Supervisors:
Alon Zivrin & Michal Zivan
Description:
We present an examination of facial expression detection of children in two different study environments, joint storytelling and yoga. We analyze videos of preschool children from the ENIC lab filmed during 6 months. Our data analysis combines face detection algorithms, artificial neural networks designed for emotion recognition, face recognition algorithm and image processing tools for tracking. We present results of child facial expressions during the recorded video sessions. This project was made in collaboration of ENIC-GIP labs.
Project Title:
HoloLens Face Emotion Detection
Students:
Ilya Smirnov, Dani Ansheles, Denis Turov
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
The project will represent face emotion detection algorithm, that runs on Microsoft HoloLens device.
The purpose of the project is providing to User real-time detected emotion as attached smiley near visible faces
The project could be used to help people, with inability to identify and describe emotions by themselves. As a result, people suffering from this disorder will be able to understand other people emotions by attached smiley.
Project Title:
An Extrapolated Dynamic String Averaging Method
Students:
Supervisors:
Simeon Reich and Rafał Zalas
Description:
The project deals with projection methods which have their origin in computerized tomography and image processing. These methods solve an optimization problem which seeks a point in the intersection of convex (constraint) sets. They do so by an iterative application of operators to the approximation point. We present a certain acceleration technique called "extrapolation". This technique allows the algorithms to take bigger steps and accelerates their convergence. We show theoretical results for the convergence of methods considered in the string averaging framework. In addition, we present a library that was developed for MATLAB which may allow researchers in the field to devise and experiment with different projection methods.
Project Title:
Students:
Supervisors:
Alon Zvirin and Yaron Honen
Description:
Developing effective unsupervised learning techniques is an essential stepping stone towards next generation machine learning models. Such models would no longer be bottlenecked by their dependence on massive labeled datasets which are often difficult or impossible to obtain. We propose a novel architecture for deep feature extraction from unlabeled data and intelligent labeling of data using an implicitly defined and learned symbolic language. The model can then be used in a semi-supervised context to reduce the amount of labeled data necessary for training.
Project Title:
Augmented Emergency Response
Students:
Avi Kotlicky,Arik Rinberg
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
We developed a 3D Augmented Reality Multi-User emergency response application. Our application allows all users to simultaneously see the same scene, while a command center on a PC sees in real time users and their movements displayed on a real time map.
The command center can communicate with all the users in scene, display hologram warnings in areas and manage all the users. The users may send warnings and request help from the command center and other users. The users also have a user location system that allows them to keep track of their team.
Project Title:
Students:
Sami Abdu and Liza Rakhlina
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
For many architects, the biggest challenge is often giving the client a clear understanding of how the design will look in real life.
The idea behind the project is to visualize the Interior and the Exterior design of an architecture model in VR, so that clients could gain an understanding of how a design will look to scale in a fully 3D environment.
Project Title:
Students:
Supervisors:
Yaron Honen and Boaz Sternfeld
Description:
HoloMessages is an Augmented Reality system for sending and receiving office messages.
Using the Hololens device by scanning a QR code at the entrance to the office, you can load user information and leave a message.
There are new technologies that enable users to live in a new and interactive Mixed Reality world.
The goal is to provide users a simple and necessary office operation, to send and receive messages and integrate the operations into the new technological world.
Providing a much more comfortable workplace, which extend your working environment outside of your computer screen to the whole room.
Similar when you put sticky notes on your desks that make an extension of your working environment.
Project Title:
Next Generation Image Style Transfer via CNN
Students:
Supervisors:
Description:
Image Style Transfer refers to an artistic process in which a content image is modified to include the style taken from a second image. Early signs of this idea appeared in the early 2000's, but the real breakthrough came with the impressive work of Leon Gatys and his co-authors. Their method relies on pre-learned neural networks, that has been trained for image recognition. Our project focuses on expanding and improving the prime weaknesses of
Gatys’ algorithm: (i) preservation of edges; (ii) enabling treatment of large images; and (iii) speeding up the whole transfer process
Project Title:
RGB camera based heart-rate estimation using Eulerian Video Magnification (EVM)
Students:
Yonatan Amir, Doron Armon, Neriya Mazzuz
Supervisors:
Ron Slossberg and Gil Shamai
Description:
The purpose of our project is to extract information and relevant medical indicators of a potential patient using an RGB camera.
We implemented few techniques in order to extract pulse in real time using a standard web camera based on a number of researches.
The main research is "Eulerian Video Magnification for Revealing Subtle Changes in the World", which published at MIT.
In addition, we have created a visualization the subtle changes in the patient's facial blood flow.
Project Title:
Real-Time 3D Face Reconstruction
Students:
Supervisors:
Description:
Fast and robust three-dimensional reconstruction of facial geometric structure from a single image is a challenging task with numerous applications.
In this project, I implemented and experimented with the iterative CNN models for extracting geometric structure from a single image.
The next step in the project was to apply the model first to videos, and from there to optimize the system for real-time 3D face reconstruction using a webcam.
Project Title:
Deep Learning of Compressed Sensing Operators with Structural Similarity Loss
Students:
Supervisors:
Description:
Compressed sensing (CS) is a signal processing framework for efficiently reconstructing a signal from a small number of measurements, obtained by linear projections of the signal. In this paper we present an end-to-end deep learning approach for CS, in which a fully-connected network performs both the linear sensing and non-linear reconstruction stages. During the training phase, the sensing matrix and the non-linear reconstruction operator are jointly optimized using Structural similarity index (SSIM) as loss rather than the standard Mean Squared Error (MSE) loss. We compare the proposed approach with state-of-the-art in terms of reconstruction quality under both losses, i.e. SSIM score and MSE score.
Source code available here: https://github.com/yochaiz/SSIM
Project Title:
Students:
hadar ben efraim, dror levy
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
We developed a VR game that based on the popular game "Fruit Ninja". Our application simulates several environments. The player's main goal is to win as much as points as he can, by cutting fruits and avoiding different obstacles. We added few more functionalities to the game so the player can enjoy both the game and the VR experience. The application gives the player an extraordinary experience that enables him to use several senses in order to get a high score.
Project Title:
FindIt - Object Detection and Tracking on HoloLens
Students:
Supervisors:
Yaron Honen and Boaz Sternfeld
Description:
FindIt is an HoloLens application for finding objects in home environment. The idea is to help us find missing stuff at home by tracking the objects around us and remember their locations. The app also tracks the objects as they change their location in the room.
The app has 3 modes:
- Scan – scan the room to find objects. This mode's purpose is to initialize the app's knowledge about the objects the user wants to track. The app is taking photos which will later be processed by Object Detection Model to detect the objects in the room.
- View – view the scan results. In this mode, the user can see the detected objects and their location.
- Find – find the objects and track changes. This mode is the main usage of our app. When the user says "Find -Object Name-", a box will appear at the exact same place as the real object. Also, the navigation system will guide the user to its location.
Project Title:
Students:
Alex Salevich, Edi Frankel
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
Our project's goal is to overcome this challenge and present a new way of interacting with an application via recognition of motion and shape instead of menus and buttons.
Our solution will allow to link the controller motion in 3d space to a certain behaviour in the application.
In our case we will implement the idea as a game in which the user plays a wizard that can draw his spells as shapes and for each one of them a predefined spell will be activated.
Project Title:
Students:
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
The old version of this project was a stand alone application with fixed models
and no control over the presentation of the holograms.
The goal of this project was to extend the application in 2 ways:
1) The app will have a user interface for user control over the presentation at computer which the app is running at.
2) The app will act as a server that can be connect to by clients so that they control the many features of the main application. In order for this to work a second Client side application has been developed as a part of this project.
Project Title:
Students:
Dor Granat , Ran Yehezkel
Supervisors:
Description:
The purpose of our project was to create an application that can detect faces and estimate their pose and expressions.
At real-time, we can identify faces and their poses in a video, and render different 3D models of faces with various expressions over the original faces.
Project Title:
Faster and Lighter Online Sparse Dictionary Learning
Students:
Shay Ben-Assayag & Omer Dahary
Supervisors:
Description:
Sparse representation has shown to be a very powerful model for real world signals, and has enabled the development of applications with notable performance. Combined with the ability to learn a dictionary from signal examples, sparsity-inspired algorithms are often achieving state-of-the-art results in a wide variety of tasks. However, most existing methods are restricted to small dimensions, mainly due to the computational constraints that the dictionary learning problem entails. In the context of image processing, this implies handling small image patches instead of the entire image. A novel work which has circumvented this problem is the recently proposed Trainlets framework, where the authors proposed the Online Sparse Dictionary Learning (OSDL) algorithm that is able to efficiently handle bigger dimensions. This approach is based on a double sparsity model which uses a new cropped Wavelet decomposition as the base dictionary, and an adaptive dictionary learned from examples by employing Stochastic Gradient Descent (SGD) ideas.
In continuation to this work, which has shown that dictionary learning can be up-scaled to tackle a new level of signal dimensions, our project is focused on studying and improving OSDL. In this report, we present several modifications to the algorithm which are aimed at dealing with its limitations, results of experiments conducted on high dimensional large datasets, our conclusions and suggestions for future work.
Project Title:
Students:
Natan Yellin, Ay Kay, Dima Trushin
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
We developed a 3D Augmented Reality Multi-User application. Our application allows all users to simultaneously see the same scene at different angles. The Users are able to cooperatively manipulate objects and plan events in the scene.
Project Title:
Students:
Roi Glink, Shay Michali, Elad Alon
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
We developed an Oculus Android gearVR application with both entertainment and educational purposes. Our application mainly offer an AR gameplay while still integrating some VR scenes and thus creating a Mixed reality experience.
The user will be able to stroll at any random location with the headset and with minimal preparation of specific set of images and key points will visit an “AR Museumâ€. That experience is gained by our application which replaces these images with art paintings, therefore offering various options for the user to experience in one room.
Additionally, the application offers both VR and AR interactions with the user. The user can control the navigation of scenes and the content of the museum to his points of interest.
For Example - the user can choose to see Van-Goh paintings and the museum will be all Van-Goh. Afterwards with only one click the user will be able to see Asher paintings and so on.
Project Title:
Image Inpainting Using Pre-trained Classication-CNN
Students:
Supervisors:
Description:
Image inpainting is an extensively studied problem in image processing, and various tools have been brought to serve it over the years. Recently, effective solutions to this problem based on deep-learning have been added to this impressive list. This paper offers a novel and unconventional solution to the image inpainting problem, still in the context of deep-learning. As opposed to a direct solution of training a CNN to fill-in missing parts in images, this work promotes a solution based on pre-trained classification-oriented CNN. The proposed algorithm is based on the assumption that such CNN's have memorized the visual information they operate upon, and this can be leveraged for our inpainting task. The main theme in the proposed solution is the formulation of the problem as an energy-minimization task in which the missing pixels in the input image are the unknowns. This minimization aims to reduce the distance between the true image label and the one resulting from the network operating on the completed image. A critical observation in our work is the fact that for better inpainting performance, the pre-training of the CNN should be applied on small portions of images (patches), rather than the complete images. This ensures that the network assimilates small details in the data, which are crucial for the inpainting needs. We demonstrate the success of this algorithm on two datasets: MNIST digits and face images (Extended Yale B), showing in both the tendency of this method to operate very well.
Project Title:
Students:
Supervisors:
Boaz Sterenfeld and Yaron Honen
Description:
As part of the project I created holograms.
I built a prism to redirect ray of light, and wrote code to transform inputs to the output format that can be reflected by the prism.
The input can be either a 3D model or an input from a webcam
Project Title:
Quad-copter Remote Control
Students:
Supervisors:
Description:
In this project, our goal is to give the user a simple control scheme to fly a drone. The possible commands are take-off, hover and land.
As we will demonstrate, we were able to control our UAV using a small, affordable and easy-to-operate remote controller.
Also, using the Raspberry Pi as our computing module, we have created a portable solution that can be mounted on the drone itself.
Project Title:
Multi AR - Real-Time Augmented Reality Multiplayer
Students:
Michael Pekel, Ofir Elmakias
Supervisors:
Matan Sela and Boaz Sterenfeld & Yaron Honen
Description:
An augmented reality real-time multiplayer where the player is literally part of the game - he can fight creatures or other real players, do quests together and interact indoor and outdoor.
The project was based on the Google Tango platform - acquiring data about the environment (walls, planes) in real time and a precise (less than 1 meter) relative position.
The game mechanism includes:
‣ Full multiplayer ecosystem - client\server udp communication.
‣ Physics - walls/planes interaction.
‣ Shoot-Hit mechanism - the real players collider position is calculated in real-time as long as the bullet trajectory.
‣ Single player quests + collaborated puzzle quests.
‣ Integration with Google Daydream controller.
‣ Single & stereo ("dual screen VR") modes.
Many thanks for our supervisors & supporters: Yaron Honen, Boaz Sternfeld, Matan Sela & Alexander Porotskiy.
The demo clip could be found at: https://www.youtube.com/watch?v=EHyYArZsYlA
Project Title:
Identification of Swarm Members
Students:
Supervisors:
Description:
This project handles the identification of a specific object instance in a swarm of similar objects. We have created an image processing component that recognizes the existence of peers in a given frame- with a main goal of identifying peer drones.
Project Title:
Hybrid Video Coding at High Bit-Rates
Students:
Supervisors:
Description:
In this project we explore the prevalent hybrid video-coding concept that joins transform-coding and motion-compensation. Specifically, we study the necessity of transforming the motion-compensation prediction residuals for their coding at high bit-rates.
Our research relies on empirical statistics from a simplified motion-compensation procedure implemented in Matlab, and also from the reference software of the state-of-the-art HEVC standard. Our results show that the correlation among the motion-compensation residuals gets lower as the bit-rate increases, supporting the marginal use of transform coding at high bit-rates (i.e., the residuals are directly quantized).
We also developed a research tool that provides data from intermediate stages of the HEVC. The data mainly include motion-compensation residuals, motion vector data, block and frame types, and bit-budget of components. It is formatted in a structure suitable for easy usage in Matlab for future research projects.
Project Title:
Perlin City - Procedural 3D City Generation Project
Students:
Supervisors:
Yaron Honen and Boaz Sternfeld
Description:
Based on a known method of creating a single random city block using perlin noise, we have created a method of procedurally generating an infinite city without the intervention of human designer.
Perline city is deterministic, realistic and beautiful.
To achieve best performance possible, we used object pooling, separation of building objects upon multiple frames using coroutines, detecting when to create and destroy and addition of detailed objects that are displayed and hidden based on distance
Download Demo:
https://drive.google.com/open?id=0B8C-GqYShcqub1pINW0zNk5obHc
Project Title:
Students:
Supervisors:
Daniel Raviv and Yaron Honen and Boaz Sternfeld
Description:
This is a virtual reality simulation of a vehicle with a camera that uses visual looming cue to navigate and avoid obstacles.
We visualized the looming in different ways and compared 2 methods for looming calculation.
One using ranges between the camera and the obstacles, and the other using the temporal change of the texture density of the obstacle.
Project Title:
Rate ditorsion optimized tree structures for image compression
Students:
Moshiko Elisof, Sefi Fuchs
Supervisors:
Description:
1D and 2D Signals compression using improved tree coding, exploiting similarities of dyadic blocks by merging them into one tree leaf, allowing for adaptive tree-structure.
The studied algorithms compensate for an inherent issue in standard tree-based signal coding - squared blocks which limit the ability to reduce representation bit-cost.
Based on:
[1]Rate-Distortion Optimized Tree-Structured Compression Algorithms for Piecewise Polynomial Images by Rahul Shukla, Member, IEEE, Pier Luigi Dragotti, Member, IEEE, Minh N. Do, and Martin Vetterli, Fellow, IEEE .
[2]Image compression via improved quadtree decomposition algorithms by E. Shusterman and M. Feder
Project Title:
CutOutRL - Visualizing Neural Networks with Scribbles
Students:
Yonatan Zarecki ,Ziv Izhar
Supervisors:
Description:
Deep neural networks (DNNs) have been very successful in recent years, achieving state-of-the-art results in a wide range of domains, such as voice recognition, image segmentation, face recognition and more. In addition, reinforcement-learning (RL) training methods combined with DNN models (deep RL) have been able to solve a wide variety of games, from PONG to Mario, purely by looking at the pixel values of the screen.
Various “games†have been proposed for challenging neural networks, testing their capacity to learn complex tasks. Some tasks are designed to give us human insight about the way the model operates.
In this project, we challenged a deep RL model with the task of segmenting an image using scribbles. We force it to achieve good segmentations by using scribble-based segmentation in a way similar to humans. We hope to gain insight on the way the network does segmentation by looking at the scribbles it generates.
Project Title:
Students:
Supervisors:
Description:
This is an open source project that intended to help people to create autonomous drone missions that operate with a pixhawk controller.
The project is written in C++ and Python in order to enable fast image processing and operating the drone in real time.
The project also includes built in missions. Our goal was to fly to a specific GPS location, scan the area for a bullseye target and land on the center of the target.
You can use the framework to create your own missions. The framework includes an API that helps to stream live video over wifi or/and record the video to file.
We’ve created this project in the Geomatric Image Processing lab at the Technion. Our goal was to create a simple framework to manage simple and complex missions represented by state machine
Read more in our github page or see the project Report and final presentation
Project Title:
Students:
Supervisors:
Omri Azenkot and Boaz Sterenfeld and Yaron Honen
Description:
VR Newsroom is an experiment in browsing online news in virtual reality. It explores ways that VR can be used to facilitate discovery and exploration of large amounts of content. Online news was chosen for the content of the experiment because of its dynamic nature. Live APIs were chosen so that every time VR Newsroom is loaded different content is displayed. Custom RSS/ATOM feeds are also supported.
Project Title:
Virtual Reality interior designer
Students:
Frenkel Eduard and Salevich Alexander
Supervisors:
Mata Sela and Yaron Honen
Description:
Our project's goal is to allow a user to experience a sense of presence in a room while designing it.
We intend to build a designing tool using virtual reality which will allow to design a room space while being present inside a virtual room;
The tool will allow the user to design a room interior and be able to experience his work product at the same time.
Our solution will allow to design better suited environment for the customer by letting the designer the ability to show the customer the final design and get his feedback.
Project Title:
DeepFlowers - Online Flower Recognition using Deep Neural Networks
Students:
Yonatan Zarecki and Ziv Izhar
Supervisors:
Description:
These days, it seems like everyone has their own smartphone and that
internet connection is available everywhere even in the most distant
corners of nature reserves. A challenging task for nature lovers is the
task of flower recognition, even with heavy big and heavy flower guide
books it is hard to identify each flower species exactly, and for
amateurs finding anything is using these guides can be a monumental task
by itself, Differences between flowers species can be very subtle, and
not easy to detect even for an expert's eye. Another challenge a flower
classifier has to face is the sheer amount of flowers in the world, or
in a specific country.
In this project we try and harvest the
power of deep convolutional neural networks (CNNs) for our recognition
task which have proven to be successful in similar tasks, and using data
given to us by Prof. Avi Shmida of the Hebrew University, build a
flower recognizer with an open online API for all to use.
Project Title:
Students:
Tom palny, Shani Levi and Nurit Devir
Supervisors:
Yaron Honen and Boaz Sternfeld and Omri Azencot and Hagai Tzafrir
Description:
Our system consists of three main parts:
The first one is to receive a 3-dimensional matrix represents the MRI scan of the brain. We used Matlab in order to create 2-dimensional images from the given matrix. Each image was saved as a PNG file and represent a specific slice of the brain.
The second part is to load the images into Unity and create a 3-dimensional object from them. In Order to build this object, we used the Ray Marching algorithm.
The last part was to implement the ability to present the object in virtual reality using HTC vive and allow features which will give the user the feeling of the 3-dimensional object.
Our project supports the following features:
• Rotating – rotate the brain using the handheld controller.
• Cutting – Cut the brain along the three X, Y and Z axes.
• Zoom – zoom in and zoom out.
• Reset button – turn back to the initial model.
• Masking – emphasize different parts of the brain according to the user choice: The user can choose one of the two options representing the parts he wants – color it or remove the unwanted parts and see only the wanted one. After using the presentation way, the user can choose the specific part in which he is interested.
Project Title:
Students:
Supervisors:
Description:
In recent years, the introduction of deep reinforcement learning allowed rapid progress in the pursuit after implementing general AI. One of the long-standing challenges withholding further progress is designing an agent that operates in a hierarchical manner with temporal abstractions over its actions. We present a system which disassembles the learning into multiple sub-skills without external assistance. The system consists of a deep recurrent network which learns to generate action sequences from raw pixels alone, and implicitly learns structure over those sequences. We test the model on a complex 3D first person shooter game environment to demonstrate its effectiveness.
Project Title:
Students:
Ksenia Kaganer , Dima Trushinand Adi Mesika
Supervisors:
Boaz Sternfeld and Yaron Honen
Description:
We developed a 3D Anatomic learning application.
Our application assist you in the learning process by creating a realistic Virtual Reality environment.
You can explore all the human body parts in a very detailed level.
Navigate between different body layers, e.g. skin, muscles, bones, internal organs etc. and see all the terminology names of each body part.
In addition, you can walk around the body naturally, have a look at the body from every aspect you want and holding a VR plane that slice the body and get a different anatomic cuts.
Project Title:
Students:
Ksenia Kaganer,Eran Tzabar
Supervisors:
yaron honen and Avi Parush and Maayan Efrat
Description:
It is widely recognized that in celiac disease, to learn and adhere to
the gluten free diet is essential for ensuring a good quality of life.
It is important that the education process adopt strategies to motivate
and make the learning effective, particularly for children and
adolescent.
In this context, new technologies can help make the learning process more engaging.
The main idea is to use a game approach in order to make the learning and training process more engaging and intuitive.
Project Title:
Puppify - Automatic Generation of Planar Marionettes from Frontal Images
Students:
Elad Richardson and Gil Ben-Shachar
Supervisors:
Anastasia Dubrovina and Aaron Weltzer
Description:
In this project, we propose a method for fully automating the body segmentation process, thus enabling a wide variety of consumer and security applications and removing the friction caused by manual input. The process starts with a deep convolutional network, used to localize body joints, which are refined and stabilized using Reverse Ensembling and skin tone cues. The skeletal pose model is then exploited to create "auto-scribbles": automatically generated foreground/background scribble masks that can be used as inputs for a wide range of segmentation algorithms to directly extract the subject's body from the background. Simple segmentation aware cropping produces individual body part crops which can be used to generate a planar marionette for repositioning and animation.
Project Title:
Students:
Sapir Eltanani and Simona Gluzman
Supervisors:
Description:
3D game which relives the experience
of playing the old well known game 'Snake', but this time in a three
dimentional world in VR. The game is developed for the leap motion and
oculus rift devices.
The aim remains the same, the player has to
eat the most of the flying foods to earn points. Every eaten food grows
the snake longer, but once the player touches the objects in the world
around him or the snake's body he loses.
Project Title:
Imae Segmentation Using Multi-Region Active Contours
Students:
Chen Shapira and Tamir Segev
Supervisors:
Description:
Our project focused on the Multi-region Active Contours with a single Level-Set function
method. This method allows quick & accurate image segmentation on 2d and 3d
images. This is done by dividing the image into multiple regions by calculating a single
nonnegative distance function, which is easily extracted using the Voronoi Implicit Interface
Method.
Project Title:
Augmented Reality in Road Navigation
Students:
Supervisors:
Description:
In this work, we propose a vision-based solution for globally localizing
vehicles on the road using a single on-board camera and exploiting the
availability of priorly geo-tagged street view images from the
surrounding environment together with their associated local point
clouds. Our approach is focused on the integration of image-based
localization into a tracking and mapping system in order to provide
accurate and globally-registered 6DoF tracking of the vehicle’s position
at all times
Project Title:
Students:
Noam Yogev, Roee Mazor, Efi Shtain, Vasily Vitchevsky, Sergey buh, Alex Bogachenko and Daniel Joseph
Supervisors:
Description:
We built a platform to be used to create
autonomous indoor flight capable drones and implemented a system based
on computer vision to utilize and demonstrate this platform by
performing four tasks, as described by a national contest sponsored by
the Pearls of Wisdom voluntary association. Currently, such platforms
are researched from commercial and academic points of view, but no
finished products have been released.
The computer vision part of the project is responsible for
communicating with the quadcopter’s navigation module, in order to guide
the quadcopter to the next target, identify key objects and trigger the
execution of various required auxiliary actions.
The software will run on an ARM-based companion computer running
Linux, mounted on the drone. It will receive images from one or multiple
high speed cameras on the quadcopter and ROS messages from the
navigation module running on the same computer. The OpenCV library will
be used to implement the required functionalities.
Project Title:
Supervisors:
Description:
The need for real-time 3D reconstruction is becoming more and more
apparent in today's world. Depth Sensors are being marketed today in
consumer laptops and tablets. In the near future we expect an increase
in availability of mobile devices with depth sensors, and therefore also
a need for highly efficient real-time 3D reconstruction methods. Our
project's goal is to enable these devices to preform 3D reconstruction
in real-time. Our solution uses the input from a moving depth sensor to
estimate the camera position and build a 3D model. The implementation
harnesses the GPU to achieve real time preformace while taking into
account the limitations of mobile devices and putting a strong emphasis
on optimizations throughout the pipeline.
Project Title:
Automatic 3D Face Printing
Students:
Hila Ben-Moshe and David Gelbendorf
Supervisors:
Description:
Developing automatic process for building a 3D face model from a
GIP facial video file. We first use Viola Jones Face Detection
Algorithm to detect which frame should we choose from the video.
Detection of face features and the movement of face are taken under
consideration when choosing the best frame automatically. Than we
process the selected frame, including automatic choose of bounding box
and with fixing missing eyebrows. Finally, we write the desired model to
STL format, a known format to print in 3D.
Project Title:
Image Segmentation and Matting in Realtime on a Mobile Device
Students:
Supervisors:
Description:
In the our project we've implemented a scribble-based algorithm for
extracting object from natural photos and pasting them seamlessly into a
different background under the constrains of a mobile device
computational power.
The algorithm was first developed and tested on a personal computer with
the help of openCV’s C++ libraries and was then ported to Android using
the Native Development Kit.
We used the Android Software Development Kit in order to wrap the
algorithm in a user friendly interface to create an application that
anybody can use.
Project Title:
Rigid ICP Registration with Kinnect
Students:
Choukroun Yoni and Semmel Elie
Supervisors:
Description:
The main goal of the first part of the project was to perform an
Iterative Closest Point registration on two depth maps obtained using
the Kinect depth sensor in C++ on the windows platform.
The other purposes of this first part was to learn how to integrate
alone big libraries (dynamic or not) to the project and to handle with
the difficulties of implementing an algorithm on the different classes
of the libraries whom do not match necessarily one with the other.
The second part of the project was to bond, two by two with the
precedent algorithm, different scan frames get by the Kinect with the
help of its motor to get a whole body depth image.
Project Title:
Students:
Supervisors:
Alon Zvirin and Yaron Honen
Description:
בניית ארגז כלים לאנליזה ראשונית של מודל\משטח פנים תלת-מימדי בניסויים קליניים עבור הרופאים. ארגז הכלים יאפשר לרופאים בין היתר להציג מודל פנים מיטבי של הנבדק לאחר סריקת וידאו של פניו, למצוא נקודות עניין על גבי המודל\משטח ולמדוד מרחקים גאודטים בין הנקודות השונות לפי שיקול דעתו של הרופא. האפליקציה מקבלת קובץ וידאו תלת מימדי ממצלמת גיפ ויודעת באופן אוטומטי לבחור את תמונת העומק הטובה ביותר מתוך סרט הוידאו. בתמונת העומק יבחרו באופן אוטומטי שלוש נקודות במרכז הפנים בעזרת ויולה וג'ונס. בעזרת ASM ימצאו באופן אוטומטי 68 נקודות בכל משטח הפנים המוצג. אלג' שפותח ע"י הסטודנט יודע באופן אוטומטי לחלוטין לבחור מתוך כל הנקודות 12 נקודות מרכזיות (5 בפה, 4 בעיניים, 3 באף). באמצעות הממשק שפותח ניתן לראות את המסלולים על גבי הפנים ,בין הנקודות הנבחרות ולקבל באופן מיידי את המרחק הגאודטי ביניהן (fast marching). בנוסף, ניתן לשמור את כל הנתונים (נקודות,מרחקים,מס' תמונה, הערות וכו') לקובץ אקסל ולעלות אותם במועד מאוחר יותר.
Project Title:
Printed circuit boards detection and image analysis
Students:
Giorgio Tabarani and Roi Divon
Supervisors:
Description:
In recent years and with recent events in the country, arose the
demand to be able to identify printed circuit boards taken by the
Israeli Police in crime scenes in order to connect between cases and
identify the source of these boards hoping to avoid similiar and
unpleasant incidents in the future.
The police takes snapshots of printed circuit boards from every
crime scene, mostly distorted circuits due to burns or fractures, and
tries to identify their origin from manual inspection and by guessing.
In this project we were asked to develop a basic system which can
perform the aforementioned identification automatically given a picture
of a shred and a database of pictures of circuit boards which usually
appear in crime scenes.
Project Title:
Generating 3D Colored Face Model Using a Kinect Camera
Students:
Rotem Mordoch and Nadine Toledano and Ori Ziskind
Supervisors:
Matan Sela and Yaron Honen
Description:
The constant development of cheap depth cameras, together with the
ongoing integration of them on mobile devices, offers the potential of
many new and exciting applications covering various of different fields.
This includes personal everyday use, commercial objectives and medical
solutions. In our project we propose a system which allows the user to
easily create a colored 3D facial model of its own. The objective of
this project is to build a user-friendly system for generating a 3D
colored facial model. The solution we offer combines open source
techniques for face detection in an image and a 3D reconstruction
algorithm. We integrate these techniques to create a common algorithm
which produces our goal. The system we have built uses depth camera
stream to capture a subject’s face on each frame, and uses this
information to generate a high quality colored 3D facial model. We
demonstrate our results and optimizations to the solution, and offer
possible future opportunities to continue our work.
Project Title:
Solving Simultaneous Linear Equations Using GPU
Students:
Oriel Rosen and Haviv Cohen
Supervisors:
Description:
Image processing tends to demand highly complicated computations.
Though some programming languages (as Matlab) are very comfortable for mathematical usage,
they are less than ideal in terms of performance, leading to programs which run far too much time.
The solution to that problem is to use “stronger†tools at the choke-points of the computation.
By programming with low-level languages and by using parallelism,
we can drastically improve our program’s performance.
Project Title:
Freehand Voxel Carving Scanning on a Mobile Device
Students:
Supervisors:
Description:
3D scanners are growing in their popularity as many new
applications and products are becoming a commodity. These applications
are normally tethered to a computer and/or require expensive and
specialized hardware. Our goal is to provide a 3D scanner which uses
only a mobile phone with a camera. We consider the problem of computing
the 3D shape of an unknown, arbitrarily shaped scene from multiple color
photographs taken at known but arbitrarily distributed viewpoints using
a mobile device. The estimated camera orientation and position in 3D
space obtained from publicly available SLAM libraries permits us to
perform a 3D reconstruction of the observed objects. We demonstrate that
it is possible to achieve a good 3D reconstruction on a mobile device.
Project Title:
3D Image Fusion using ICP
Students:
Supervisors:
Alon Zvirin and Guy Rosman and Yaron Honen
Description:
This project deals with the ICP
algorithm and uses it to create a complete three-dimensional model of a
rigid object. First, a wrapper for an ICP algorithm included in PCL
library which fuses 3D images taken by GIP Technion laboratory camera
was written, and its running parameters were optimized. Afterwards,
various improvements were implemented for an ICP algorithm using
distance function of point to surface, parameters were examined, and
fusion algorithm was written. Finally, the program was integrated to the
GUI of GIP Technion laboratory
.
Project Title:
Students:
Shahar Sagiv, Omri Panizel, Loui Diab and Waseem Ghraye
Supervisors:
Description:
We created a new type of game, which combines the competitive aspect of
the Bingo game with the great fun of solving a puzzle. This game is
being played simultaneously between 4 players (on 4 different devices)
who compete by solving the given puzzle. Every player can see the real
time progress of the other players with a map showing their boards.
Every Puzzle has a title name and the user have a choice to stop solving
the puzzle when he recognize the image, and try to guess the image
title. This gives the game educational value, as the players learn to
recognize places, animals, and celebrities.
Project Title:
Students:
Anna Ufliand and Sergey Yusufov
Supervisors:
Description:
המטרה העיקרית של הפרויקט היתה פיתוח מערכת אשר תהיה מסוגלת להתתממשק עם מכשיר לבדיקות רפואיות, אשר פותח בפקולטה להנדסה ביו-רפואית, לקרוא ולהציג את הנתונים מהמכשיר הנ"ל בצורה נוחה, יעילה וכך שיהיה ניתן לגשת לנתונים האלא ממכשירים שונים.כאחד הקריטריונים החשובים עבורינו היה לפתח מערכת אשר תאפשר גמישות מירבית. למכשיר ביו-רפואי יכולות להיות הרבה קונפיגורציות שונות, זו בעצם פלטפורמה המאפשרת הרכבת חיישנים שונים לביצוע בדיקות שונות. היה לנו חשוב להתייחס לכך ולפתח מערכת אשר לא תגביל את סוג וכמות החיישנים אשר בעזרתם מבצעים בדיקות.
Project Title:
Students:
Supervisors:
Description:
Artists, designers and architects use imbalance to their advantage to produce surprising and elegant designs.
The balancing process is challenging when manipulating geometry in a 3D modeling software,
since volumes are only represented by their boundaries.
Our goal is to modify volume shape such that once printed, the model stands.
To do so, we manipulate the inner voids and sometimes it will not be enough and we will also consider deforming the model.
These two manipulations change the mass distribution and thus the center of mass position.
Project Title:
Solving Classification Problem on Hyperspectral Images
Students:
Talor Abramovich and Oz Gavrielov
Supervisors:
Description:
Hyperspectral Imaging is a spectral imaging method, which includes
bands from the visible light as well as infra red. Unlike the 2D color
images, which only use red, green and blue, hyperspectral image includes
a third dimension of spectrum. This information can be used to classify
the objects in the image, and to define the difference between asphalt,
plants and water. It could even show the difference between real leaf
and a plastic one. In our project we used several classification
algorithms, including KNN, PCA and KSVD to classify four hyperspectral
images. We compared the results and found which algorithm gives the best
classification and which is the most efficient..
Project Title:
Students:
Supervisors:
Description:
משחק אינטראקטיבי בין שחקנים במהלכו כל שחקן יכול לשלוח לשחקן אחר תמונה, השחקן השני צריך לנחש מה יש בתמונה לאחר שהופעל אלגוריתם כלשהו לעיבוד תמונה המקשה על הזיהוי. השחקנים מקבלים ניקוד כתלות בזמן בו הצליחו לנחש. המשחק ממומש למכשירי אנדרואיד, כאשר קיים שרת ששומר את כל המידע וכל המשחקים מתבצעים דרכו.
Project Title:
Descriptors-Based Stereo Reconstruction
Students:
Anan Abu-Yousef and Anna Ginzburg
Supervisors:
Description:
In our project we built a stereo reconstruction pipeline. We compared several descriptor algorithms and checked how they work for stereo
reconstruction. We used various linear projections [1,2] to investigate the effective dimensionality of descriptors coming from stereo
reconstruction.
Project Title:
Structured Light Based 3D Reconstruction With Priors
Students:
Supervisors:
Project Title:
3D Camera for Mobile Device
Students:
Deborah Cohen & Dani Voitsechov
Supervisors:
Project Title:
Touch tic-tac-toe with depth camera
Students:
חוסיין עותמאן ועבדל מסיח מבאריקי
Supervisors:
Description:
מטרת הפרויקט היא מימוש משחק איקס עיגול מגע בשימוש ב ומקרן ומצלמת . משחק איקס עיגול הינו משחק קלאסי לשני משתתפים שבו נתון לוח בגודל שלוש על שלוש. ומשתמש אחד מסמן והשני מסמן לסירוגין. מטרת כל משתמש היא ליצור רצף והראשון שמצליח ליצור רצף מנצח. מטרת פרוייקט זה הינה לממש גרסא של משחק זה על לוח מגע ללא מסך מגע. מקרן יקרין את המשחק על לוח לבן ואז המשתתפים יסמנו את המיקום שבו הם רוצים לשים את סימונם. זיהוי התנועה והסימונים נעשה בעזרת מצלמת
Project Title:
Quaternion K-SVD for Color Image Denoising
Students:
Supervisors:
Description:
In this work, we introduce the use of Quaternions within the field of sparse and redundant representations. The Quaternion space is an extension of the complex space, where each element is composed of four parts – a real-part and three imaginary parts. The major difference between Quaternion space H and the complex C space is that the Quaternion space has non-commutative multiplication. We design and implement Quaternion variants of state-of-the-art algorithms OMP and KSVD. We show various results, previously established only for the real or complex spaces, and use them to devise the Quaternion K-SVD algorithm, nicknamed QK-SVD.
Project Title:
Structured Light Based 3D Reconstruction with Priors
Students:
Itamar Talmi and Ofir Haviv
Supervisors:
Description:
פרויקט זה בא לבחון שימוש באלגוריתם PCA לצורך פתיחת נעילת מכשיר אנדרואיד על ידי זיהוי פנים. פרויקט זה מדגים כיצד ניתן, ע"י למידה של פני האדם כלשהו בתאורות שונות, בניית בסיס PCA של אותו אדם, ובעזרת בסיס זה, לאמת מישהו שמנסה להזדהות בעת פתיחת נעילה של מכשיר אנדרואיד.
Project Title:
CamPong – Smartphone PONG using Camera and Built-in Projector
Students:
Nofar Carmeli and Rom Herskovitz
Supervisors:
Description:
In this project, we introduce the use of a mobile phone, equipped with a
camera and a projector, to allow real time hand detection. We present a
demo of an interactive pong game that is controlled by the players’
natural hand movements during the game. To the best of our knowledge
this is the first known use of a commodity cellular phone that uses an
inbuilt projector to perform real time structured light projections
coupled with real time image processing.
Project Title:
Students:
Nurit Schwabsky and Vered Cohen
Supervisors:
Description:
OpenFusion is an implementation of Microsoft’s KinectFusion system. This
system enables real-time tracking and reconstruction of a 3D scene
using a depth sensor. A stream of depth images is received from the
camera and compared to the model built so far using the Iterative
Closest Point (ICP) algorithm to track the 6DOF camera position. The
camera position is then used to integrate the new depth images into the
growing volumetric model, resulting in an accurate and robust
reconstruction. The reconstructed model is adapted according to dynamic
changes in the scene without losing accuracy.
Project Title:
3D Stereo Reconstruction Using iPhone Devices
Students:
Ron Slossberg and Omer Shaked
Supervisors:
Description:
Stereo Reconstruction is a common method for obtaining depth information
about a given scene using 2D images of the scene taken simultaneously
by two cameras from different views. This process is done by finding
corresponding objects which appear in both images and examining their
relative positions in the images, based on previous knowledge of the
internal parameters of each camera and the relative positions of both
cameras. This method relies on the same basic principle that enables our
eyes to perceive depth.