smellslike.ml Open in urlscan Pro
18.173.154.19  Public Scan

Submitted URL: http://smellslike.ml/
Effective URL: https://smellslike.ml/
Submission: On May 09 via api from US — Scanned from DE

Form analysis 0 forms found in the DOM

Text Content

Smells Like ML

 * Archives
 * Tags


HEY, WE'RE TERRY & SALMA 👋

The technical tagteam behind this blog. We aim to showcase experiments and
innovations in Artificial Intelligence and applied Machine Learning.




THIS IS THE REMYX

Check out remyx.ai Why do we care about ML in product? As you’re reading this,
one might assume that you are familiar with machine learning (ML) concepts. But
since we aim to reach a broad, ML-curious developer audience with this post, let
us simply frame our interest in ML-based product features which help us to make
decisions using cheaper data. The case for custom ML Many practical machine
learning tasks come down to differentiating between only a few object categories
in a constrained environment....

January 11, 2023 · 3 min · Terry Rodriguez & Salma Mayorquin


FILMGEEKS 3

Check out the FilmGeeks3 Collection In last year’s post on generative models, we
showcased theatrical posters synthesized with GANs and diffusion models. Since
that time, Latent Diffusion has gained popularity for faster training while
allowing the output to be conditioned on text and images. This variant performs
diffusion in the embedding space after encoding text or image inputs with CLIP.
By allowing the generated output to be conditioned on text and/or image inputs,
the user has much more influence on the results....

October 21, 2022 · 3 min · Terry Rodriguez & Salma Mayorquin


ROS NEURALRECON

Check out the ros_neuralrecon repo here! About a month ago, we remarked on some
exciting modifications to NeuralRecon, which can generate dense scene
reconstructions in real-time using TSDF fusion on posed monocular video.
Specifically, we noted that the image feature extraction could be shifted to the
edge using depthai cameras after converting the MnasMulti backbone to .blob.
Trained on ScanNet, the researchers recommend custom data capture with ARKit
using ios_logger. We found this works well if you have Xcode on a Mac and an
iphone....

October 3, 2022 · 4 min · Terry Rodriguez & Salma Mayorquin


GO NERF YOURSELF

While prototyping YogAI, our smart mirror fitness application, we dreamed of
using generative models like GANs to render realistic avatars. For the TFWorld
2.0 Challenge, we came a bit closer to that vision by demonstrating a pipeline
which quickly creates motion transfer videos. More recently, we have been
learning about reconstruction techniques and have been excited about the work
around Neural Radiance Fields (Nerf). By this method, one learns an implicit
representation of a scene from posed monocular videos....

September 26, 2022 · 2 min · Terry Rodriguez & Salma Mayorquin


CHEAPER TO FLY

In recent experiments, we’ve generated high quality reconstructions of our
apartment from video. Learning the failure modes of these methods, you will move
the camera smoothly, avoid bright lights, and focus on textured regions of the
FOV. If it all works out, you might spend more time recording video than
processing it! Automating the data collection can really reduce the cost of
mapping and reconstruction. Compared to recording from a phone/tablet, drones
move more smoothly and swiftly....

September 13, 2022 · 3 min · Terry Rodriguez & Salma Mayorquin


MEET THE FLOCKERS

In this post, we share one of our favorite “pet projects”. We first heard about
the “parrots of telegraph hill” looking for things to see in the city. But after
a couple years, we never managed to run into one of these accidentally re-wilded
parrots. Eventually, we moved to a new apartment where we could hear their
distant squawks and occassionally see a small flock of the cherry-headed
conures....

September 6, 2022 · 3 min · Terry Rodriguez & Salma Mayorquin


REAL-TIME RECONSTRUCTIONS

Modern archeologists have been able to survey larger structures more precisely
using remote sensing and photogrammetry. More recently, researchers demonstrate
applications of multi view stereo with networked embedded cameras to track
geological disturbances. In scenarios where visibility comes with high cost or
saftey risk, the ability to quickly render high-fidelity reconstructions for
offline analysis & review can be a powerful tool. Advances in techniques like
multi-view stereo and structure from motion have reduced the cost by alleviating
dependence on more expensive sensors like lidar....

August 29, 2022 · 4 min · Terry Rodriguez & Salma Mayorquin


DETECT-TRACK-LOCALIZE

In our latest experiment with Depthai’s cameras, we consider visual
localization. This relates to the simultaneous localization and mapping (SLAM)
problem that robots use to consistently localize in a known environment.
However, instead of feature matching with algorithms like ORB, we can try to
directly regress the pose of a known object. This approach uses object
detection, which is more robust to changes in illumination and perspective than
classical techniques. And without the need to generate a textured map of a new
environment, this technique can be quickly adapted to new scenes....

August 18, 2022 · 3 min · Terry Rodriguez & Salma Mayorquin


POINTCLOUD VIDEO

Lately, we’ve come to enjoy using the DepthAI OAK-D, which features an RGB
camera with stereo depth, IMU, and Intel’s MyriadX VPU. Along with this powerful
hardware combination, DepthAI provides a rich SDK to build your own embedded
vision pipelines. Many projects are included to get you started. These specs
could help bring spatial AI to the SpecMirror where we can test representing
human activities with pointcloud video. The Data First, we will generate
training samples for activity recognition models like P4Transformer....

August 14, 2022 · 2 min · Terry Rodriguez & Salma Mayorquin


ADDING VISION TO THE TURTLEBOT3

Turtlebot open-sourced designs for many low-cost, personal robot kits and the
Burger is a great platform to learn ROS and mobile robotics. The documentation
progresses from setup to more advanced robot behavior including SLAM,
navigation, autonomous driving and more! Our Build Since the turtlebot3 burger
kit doesn’t include a camera, we added the OAK-FFC-3P camera made by Luxonis
with a 100 degree Arducam to our build. This device includes deep learning
accelerators on the camera so we can reduce the compute needs on the raspberry
pi host....

April 22, 2022 · 2 min · Terry Rodriguez & Salma Mayorquin
Next Page »
© 2023 Smells Like ML · Powered by Hugo️️️ · Theme️ PaperMod