DEV Community

Maureen T'O
Maureen T'O

Posted on • Updated on

What I learned from a semester of machine learning...

Q: What have you learned recently? 💭

I recently enrolled in a undergraduate-run AI program called LearnAI, hosted by the student club Undergraduate Artificial Intelligence Group at UofT (UofT AI). The program was led by upper-year students in computer science and the course taught machine learning fundamentals and AI to students who have not yet taken the course (at the University of Toronto, there are 7 prerequisite courses to take in order to take upper-year machine learning courses). Point being, this program showed students can learn to work with machine learning/AI without needing to have all the prerequisites.

Even though I didn't understand all the math behind the algorithms and concepts they were teaching (CNN, RNN, loss function, etc.), I still managed to make it to the project phase. I'll share some details and tidbits below... 😃

Construction under way... 🚧

Amongst groups of teams, my team and I used text summarization (Natural Language Processing) to analyze a given shards of text and be able to eventually pick out keywords and combine the textual data into a coherent summary. Kind like how you don't want to read a heavy research paper, and just need to know the summary.

  • What I learned: how to find datasets, understanding RNNs/CNNs, text summarization principles, techniques to use

It was in this stage that I realized, "Holy crap, finding the right dataset and the amount of data was the hardest part!". In a larger sense, knowing the business logic and reasoning behind the purpose of our project was another big aspect we needed to consider. How would our project benefit others? What types of problems will it solve? Has there been a solution created already? To be fair, I don't think people usually understand the amount of work that goes in to machine learning research projects and the amount of energy needed to sustain them. There is so much more to selecting the right model for your project, over-fitting issues, and knowing what algorithms to use, at least for me 😄.

We first went on to Kaggle and Google Dataset and found our data set - the Daily Mail data set. After going through that repository's code to filter out what we needed, we finally got down to being able to train our models with training data. We are currently in the process!

  • What I learned: cleaning data, refactoring code, percentage allocation of training/testing data, working in a team, working with other people's code, using nltk, version control, tensorflow

Eventually, we'd like to take this idea and be able to integrate it into a software application to aid accessibility, and for note-taking purposes.

#YearEndAdventures

I recently attended the StartAI Conference 2020, the world's largest undergraduate AI conference. Guess who the keynote speaker was? None other than Professor Emeritus Geoffrey Hinton, often known as the "godfather of AI".

At his presentation,

Some psychological evidence that our visual systems impose coordinate "frames" in order to represent shapes. - StartAI Conference ft. Prof. Hinton, [Presentation slides]

How cool is that? Just like Plato's Theory of Forms, our brains compose a visual representation of objects within our visual frame to recognize them! In kinesiology, we learn that our brains build motor programs, modify them, and store them (memory). This is called a "motor representation".

Who else did we meet?

LearnAI also had the honour of hosting Sam Lightstone this week - the CTO for IBM Data, IBM Data & AI, and an IBM Fellow. Here's some fun facts I learned: Did you know that no supercomputer is more efficient than the human brain? Did you know that if you change a couple pixels in an image, imperceptible to the human eye, that it completely messes up the supercomputer?

The human brain consumes around the same Watts as a lightbulb!

After all that, we can see machine learning is both incredibly powerful but incredibly fragile to manipulation. Before this program, I hadn't had a clue of AI's capabilities nor how it really melded with other fields like healthcare (something I'm studying) and medicine. I think it's safe to say that AI is still young, yet advancing so fast that we may run out of energy to power our training models by 2040 🤯 Again, it's up to developers and machine learning researchers/institutions to start to understand how to better solve current issues as most practical machine learning applications are mainly image-recognition based, when in fact, many issues that currently hold us back reside with JSON or text-based data. Ethically speaking, how can we make AI sustainable, fun, and/or ethical?

In conclusion, the field of machine learning has grown so much and has so much more room to grow. I can't say that I'm an expert at all but this certainly intrigued me. Its applications reach far and wide, across disciplines, and ever-reaching. It's definitely worth it to see what the boundaries are for machine learning is and how we might build AI to be ethical and benefit many, not just a few.

Curious? 💡 I've listed some ML resources below:

Top comments (0)