Payal Mitra
Payal Mitra

Payal Mitra

Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities

twitterPersonal Blog
Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 30: How & where to volunteer your ML skills towards nonprofits?
Payal Mitra

Where would non profits (NPOs/NGOs) with modest to little resources benefit from Machine Learning (particularly NLP) solutions the most? A surface level answer could be 'where would they not?'. How can one find and volunteer ML skills towards such projects?

This is where my shallow (but not first) search fails. Here is a 2018 (!) McKinsey report on AI-for-social-good confirming that there should be enough problems to go around. Beyond think-pieces and AI-for-social-good bootcamps (mostly for university students), I don't easily find advertised projects tied to nonprofit missions that allow variable-commitment. Note, I am not including university initiatives or social wings of large companies (Google/Microsoft/IBM, etc).

Here are my scribbles on why there is a gap in reliable resources or 'market'places for ML volunteer work for professionals (Oops, yet another think piece).

  1. Wouldn't it be convenient to have a crowd sourced platform for such projects (an Amazon Mechanical Turk for NPO tasks? Not Kaggle). However, given the domain is cash/resource strapped, it isn't a surprise this does not exist.

  2. Was #1 not existing simply due to the lack of resources and friction to enable the marketplace? Or could it be a supply side issue? i.e., not enough professionals wanting to/being able to reliably and recurringly commit time to volunteering.

  3. Is there a demand side issue? NPOs might be wary of contracting volunteers to support their ML needs given their unsteady and uncertain stream. Further, the organisation might be fire-fighting and not be able to dedicate time to study alternatives, or may not have the ML expertise to know what is feasible.

  4. What about maintenance? Beyond a place for buyers-sellers of a need to meet, we also miss a framework that handles the aftercare of any ML solution developed by volunteers. Who takes responsibility? Many NPOs do not have the luxury of in-house tech support. What is a sustainable approach/framework to BAU?

One could of course individually contact the organisations whose missions they believe in (ChatGPT also suggests this). However, I am guilty of not following through. Or, we could build such a crowdsourced platform? I hope to have more than questions.

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 29: What computer OS is to software, CRISPR is to biotech apps, LLMs (BERTs & GPTs) are to ...?
Payal Mitra

"CRISPR gene editing technology is like an operating system that scientists can build biotech apps on" - an idea from 'The Code Breaker' by W. Isaacson that blew my mind.

CRISPR had an explosive impact on the biotech world as it offered a foundational tool that scientists/practitioners could use to bring their use-cases, or 'apps' to life. Be it to edit a gene to act as a diagnostic tool, to act as a mechanism for vaccines, or to edit out undesirable diseased genes. 'Bio-hackers' even had home-kits for gene editing!

The more ubiquitous and pervasive example that touches most humans - is how computer operating systems (OS) such as Unix-based Mac OSX, Windows, or the open-source Ubuntu OS distributions have made possible the computing devices and software that now act as an extension of our lives. (As an aside: Ubuntu philosophy)

The point I am getting to is that an Operating System acts as a platform, that unlocks the potential of human ingenuity by allowing the creation of applications (apps) on it.

In a similar way, Large Language Models (LLMs) have started resembling a platform on which other apps can be built. We are already witnessing the avalanche of new tools built atop ChatGPT and GPT-4 APIs. You can automate digital tasks with AutoGPT, code with Github Copilot, have personalised AI tutors at Khan Academy.

A nuance in this surge of LLM-powered AI apps is that these creations are built atop APIs to LLMs created by the few leading AI players, and most of them do not have actual access to the model (unless using open source models). Such mass-scale use of centralised ML models is unprecedented. Here's an Interesting twitter thread on whether LLM-owning companies will have a moat. The situation reminds me of the OS-business, and I wonder if OpenAI, etc will become the future Windows in the LLM space? Perhaps we'll also have customisable open source versions of LLMs that the community collectively maintains - much like the many flavours of Ubuntu. I hope so.

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 28: My 3 most read Newsletters for business journalism and deep thinking about society stuff
Payal Mitra

I have a bittersweet relationship with newsletters.

It is easy for the rate of collection of newsletter subscriptions/receipt of newsletter emails to far exceed the rate of absorptive reading. It reminds me of something someone said: "Buying books and reading books are two different activities/hobbies".

If you are at all interested in keeping abreast of economic/business journalism in India, or on deep think pieces on society and economics globally, try these posts:

1. Finshots ('The Ken' is a close contender)

Fascinating topics and well-researched delivery of financial business & economic news from India. Highly educative - on startup valuations, government financing, financial innovation, failures, etc. Somehow, it is free! I would definitely convert if it were paid.

The Ken is another great source on business journalism covering businesses, startups and healthcare. It is a paid subscription, and I have not fully opted in yet.

2. Balaji Srinivasan and & 'Not Boring' by Patty McMormick

I read these on and off. However, I find they both offer very useful and insightful perspectives into technology, economic and social organisation, as well as social implications. Fairly original, and well researched pieces.

3. Wait But Why by Tim Urban, Ribbonfarm by Venkatesh Rao

I am not a regular reader of these two newsletters, but they make for fun, insightful readings. 'Wait But Why' is random and fun, with topics ranging from philosophy to technology with a side of analysis of societal implications.

Re: Ribbonfarm, I like the parent blog, and thus had to subscribe to this newsletter. It is a breath of fresh air, and offers unique and thought provoking perspectives into the world around us. Sample post to win you over: 'A Big Little Idea called Legibility.'

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 27: Only a handful of labeled examples? 3 ideas from prompt-based learning research you can use to prep LLM for few-shot learning
Payal Mitra

'Prompt' has become an everyday word on account of the popularity of ChatGPT. However the idea of prompting any language model (particularly decoder, or encoder-decoder models) to get a response back in natural language has been around for a few years. Here are a few ideas from research that can help you perform few-shot learning.

This list is obviously not exhaustive. It is just indicative of some the strategies you can try if you find yourself with limited labeled data, and limited compute.

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 26: What does the LLM democratisation debate have in common with BlockChain Trilemma?
Payal Mitra

This musing is mostly born from the challenges I faced while trying to load LLMs locally (for inference and fine tuning) that are 1/ 1000th of the parameter size of GPT-4.

The dilemma around how transparent, democratised, scalable, and sustainable the hugely popular LLM capabilities (such as ChatGPT) are to practitioners across the world, reminded me faintly of the Blockchain Trilemma and Blocksize war.

Well, this is more of an analogy rather than a theoretical similarity between the dilemmas these two very different technologies face when they attempt to scale. The Blockchain Trilemma states that in any blockchain you can only fully solve 2 of the 3 problems of security, scale and decentralisation (however, there are now workarounds (yet to stablise) to solving this, such as Sharding, Roll-ups)

Granting myself a creative license to stretch the analogy. If we view the LLM deployment issues faced today from a blockchain trilemma lens, then we face -

The Blockhain Trilemma is not resolved yet, however there are promising methods the community is experimenting with, to expand the scale of transactions without compromising on decentralisation.

Can we take inspiration from the Blockchain trilemma to see how we can improve the decentralisation & democratisation of LLMs. Should they be smaller, smarter, distilled?

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 25: My self-reminders to 'Create to experience becoming' and 'without demanding it pays the bills'
Payal Mitra

I've recently been reminding myself about why I wanted to write in the first place. A daily writing habit for the self is the goal - but, publishing and expecting an external reward is counter-productive at times.

Thus, I paused and found myself going back to two pieces. The title fuses quotes from two of my favourites, Kurt Vonnegut and Elizabeth Gilbert. It contains their meditations on practicing creativity for the self, and emphasises practicing your creativity without expecting an external reward of any sort.

Sharing my fav pieces on this topic:

  1. Kurt Vonnegut's response to a letter from the children at the high school in 2006. Teaser excerpt to get you to read the original:

    • "Here’s an assignment for tonight...Write a six line poem, about anything, but rhymed.... Make it as good as you possibly can. But don’t tell, show it or recite it to anybody. Tear it up into teeny-weeny pieces, and discard them into widely separated trash receptacles. You will find that you have already been gloriously rewarded for your poem. You have experienced becoming, learned a lot more about what’s inside you, and you have made your soul grow."

  2. Elizabeth Gilbert's Big Magic.

    I had a great many takeaways, but they mostly fall into 4 buckets:

    • "do not demand creativity to pay the bills", or

    • "trust that writing loves you back, so keep showing up for it".

    • "Take care of it (your creativity), and it will take care of you."

    • "follow your curiosity"

Hope these recommendations find you, and that they are useful.

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 24: "Why AI is incredibly smart and shockingly stupid" - why I recommend Yejin's Ted Talk
Payal Mitra

"Why AI Is Incredibly Smart — and Shockingly Stupid" is a Ted Talk by AI Researcher and Professor, Yejin Choi. I highly recommend you watch this excellent 16 minute clip. It talks about the focus beyond just building ever larger and larger language models.

Should you watch this? Either you

  • are an AI researcher in either camp (scale enthusiasts, vs. let's-get-smarter-in-building-smaller-models-with-(distilled)-knowledge).

  • are looking to utilise ChatGPT-esque technology in your product. It is the user's prerogative to know the cautionary boundaries, and how to overcome them

  • are a student/AI practioner without the resources to be able to play or fine tune the models. The open source community (you're a part of it) will find a way.

  • are jumping wagons with each shiny new ML tool and are overwhelmed by the pace

Why I liked it

  • Echoes a worry I share re: making these models smaller, sustainable and more democratised? Call for/to action.

  • Excellent presentation, with insightful analogies.

  • I enjoyed Yejin's take on Common Sense Reasoning being the most important task.

  • Spells out cautionary lines for people to when integrating ChatGPT-esque tools

  • Personally, it re-emphasised the value of going long on certain notions of how you want to contribute. Think of the bigger picture and not get swayed by latest tech

Though, I do want to say - credit where credit is due. Large Language Models and techniques like Reinforcement Learning with Human Feedback are phenomenal feats, and give us fantastic starting points to distil and fine-tune task-directed models.

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 23: 'Deep Neural Networks are not interpretable' - what this means
Payal Mitra

y = b1x+b0. If x is 1D, this linear model has 2 parameters. In contrast, recent Deep Neural Network(DNN) model 'GPT-3' powering ChatGPT has 175 Billion parameters. 'GPT-4' of ChatGPT Plus has 1 Trillion. Beyond the explosion in number of parameters, DNNs learn complex non-linear functions and are said to be non-interpretable in how they learn to map input(x) to output(y).

But what does interpretability in DNNs mean? (this linked article attempts definitions). For this post, we consider a model as interpretable if we can perfectly predict the output, and can infer how input data changes affect predicted output. Is the model's decision path hard to find or is it that the theoretical framework and design choices in training the models do not easily account for interpretability?

Below I describe 3 aspects of the underlying learning theory for a large section of DNNs (excluding Bayesian Deep learning, etc) that bake in non-interpretability.

Illustration: Imagine we train a DNN to predict the agricultural crop yield of land plots. The model takes as input the raw sensor data on water levels and weather of the land.

  • Deep learning does not learn statistical relationships between dependent target variable (y) and input variables (x). DNNs do not rely on explicit input features, but learn their own 'latent' features and representations of input.
    Illustration: The crop yield prediction model does not learn explicit relations linking water levels to crop yield, but hunts for its own patterns in the sensor data (and learns its own input features) that give best performance in predicting the yield.

  • Deep learning does not explicitly account for relationships between input variables.

    Illustration: The DNN does not model how weather (e.g. excess temperature) might affect water levels. It thus can not predict how much the yield would change as a function of interlinked changes between inputs. It only predicts on observed data.

  • Most importantly, Deep learning does not assume probabilistic processes behind the data generation of either input or output.

    Illustration: Input variables weather and water levels are collected across different days and have distributions. High temperatures and low water levels might require a different crop yield prediction function, while a period with moderate rainfall follows a different yield trajectory. The likeliness or unlikeliness of the draw of a certain input event (tsunami?) should change the crop yield function. Our model does not consider the data to be generated or related in a probabilistic manner.

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 22: Wondering where statistics & statistical learning are in ML/DL?
Payal Mitra

Despite the fact that Machine Learning draws upon on aspects of Statistics and Statistical Learning Theory (SLT), it has many differences.

Much of the math, statistics, modelling and computer science algorithm concepts that make ML possible are abstracted away in present day ML/DL code frameworks (sklearn, pytorch, pytorch-lightning, etc) that allow you to train+predict using an ML model in a few lines of code. Thus, it is very natural to wonder at the differences and similarities between the concepts, particularly if you are approaching ML without academic training in/introduction to the underlying subjects.

Here I share 3 resources to hopefully clear some of the doubts.

Fundamentally the cases made above are for studying data points and trying to either infer a relation between variables, or create predictive models. In very simplistic terms:

  • Statistics -> about relationships between variables. It is based on probability theory and assumes observations are a sample drawn from a population.

  • Statistical Learning Theory(SLT) -> built on the ideas of statistics and functional analysis to make build models that can infer relationships and thus predict dependent variable given data.

  • ML -> while ML is built atop concepts from SLT, it is focussed on predicting a target variable given an input (in supervised ML), without caring about exact relationship b/w variables. It is a function approximator, and does not assume a probabilistic model for the data generation process or observed counts.

This short writeup does NOT do justice to the topic. Always happy to discuss further!

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 21: "To be everywhere is to be nowhere" Focussing on un-scattering myself
Payal Mitra

[This is a quick diary entry, that I have a feeling I'll later regret publishing. But, oh well]

If I have a theme for this year, it is to be intentional with my pursuits and energy. I used to claim to have multiple passions & aspirations, but the truth is that only some of them are my genuine passions/strengths. The others I enjoy for sampling variety/novelty/ adventure (and will continue to, in a controlled manner), or some (more professionally related) which I only do coz' I think they'll keep me shiny in the job market.

As a telling (non professional) example:

I recall a well-meaning comment 3 years ago from a friend that triggered my defensiveness. I was telling her, about the long list of sporting activities I was pursuing: Lindy hop, badminton, running, cycling, swimming. I was considering going back to beginner level kickboxing. She said something to the lines of "Payal, you need to focus. What are your goals? This is very scattered."

Similarly, at work - I was trying to shoulder too many independent responsibilities, often causing overwhelm or stress. I know I delivered the jobs very well, but I was missing out on excellence, or pure enjoyment of seeking knowledge and of creating, or the thrill of contributing to a cause I truly believed in, because I spread myself too thin.

Seneca in Letter II from 'Letters from a Stoic' [Translation by Robert Campbell] articulates well how I felt - "To be everywhere is to be nowhere".

Now, I still will pursue some activities for personal joy and no external display, as is human. But, I am selecting a few areas where I consolidate most of my energy budget.

As an example, in terms of physical pursuits - while I continue to play other sports occasionally and socially - this year is mainly about running and swimming for me. I have come to really enjoy running and am in awe of the body and mental calm. A tangible goal is my first half marathon in October this year.

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 20: Looking for something to make you enjoy the process of writing again? Read this blog post
Payal Mitra

This is not a self-help list for rekindling the joy/purpose in a craft you previously loved. Rather, it is far more tangential, and reads like a peek into the musings and of an articulate and established writer. You may take what you will from it.

The author themselves is a writer by profession, and talks about how blogging satisfies a certain itch for him. Payed subscription letters are not conducive enough an outlet for his creativity, indulgences.

The Recommended Blog: Salt Seeking by Venkatesh Rao. And here's why:

In case you are short on time, or do not fancy this piece - read the linked blog for the educative metaphor of 'salt seeking' if nothing else. It will be worth it (though it is possible you don't like it. I am reminded of a line I read that succinctly expresses this idea: "A universal applause is at least two thirds scandal").

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 19: The Supply Chain in building Machine Learning models
Payal Mitra

It recently dawned upon me that training and serving of ML models can be viewed as a supply chain across companies/part providers. This is a random musing, but if you are into the ML development, then hang on for the metaphor in the end.

Up until a few years (10?) ago, before the widespread use of cloud compute, many ML models would be trained completely in-house on-prem resources (the history of aws says AWS was launched in 2003 as a bid to manage infrastructure scaling for engineering needs). Nowadays, many in the industry use some form of cloud resources to train (/serve) their ML models salably and reliably. An analogy would be developers renting out factory equipment to perform their production deployments.

Beyond cloud compute and scaling solutions, there are several great third party tools, such as those for tracking of ML experiments, data versioning, ETL, creating pipelines from ML components, automating pipelines, and even auto-hyperparameter tuning!

The new kid on the block is 'data-centric' AI. We now have multiple players promising to provide solutions/tools to improve trained ML models by improving datasets, and not just tweaking the model architecture (Data-centric AI vs Model-centric AI).

What then is the core competency of a company that outsources several parts of the ML process? The core competency a company looking to develop an ML model in such an outsourced manner is perhaps that it owns the data, possibly the closed source code, the integration logic for various components and the right product/ market reach. We have seen similar business models where different components are outsourced, but the design, assembly, testing and improving lie with the final goods-services seller. Example: An automobile manufacturer might have different car parts sourced, but they are recognised for their ability to manufacture the car as a whole. Is that similar to the future of ML model producing entities?

A possible metaphor: companies own the data but outsource the other components before assembling all parts into one final product. This is partly like owning the soil and land, but outsourcing all other duties for tending to and growing plants to others.

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 18: Insider take: Why adoption of LLMs into core products will still be kinda slow in several industries
Payal Mitra

As everyone rushes to use OpenAI models, build atop its APIs, or use services like ChatGPT in their workstreams, many wonder if the dawn of super abundant (conversational) AI is already here.

As someone working in NLP and Deep learning in a non-MAANG or non-startup-born-in-today's-digital-era, my view is that adoption of Large Language Models (LLMs), even open sourced ones, in existing service/product industries (such as healthcare, publishing, etc), will be slower than you think

The idea of LLMs have been around for a while, most notably since BERT models in 2018, but I've seen resistance to its adoption in some companies. One might argue that the recent ChatGPT wind has woken up management professionals to the possibilities of LLMs (although I have heard too many people refer to LLMs and Generative AI and ChatGPT interchangeably - and that worries me). However the integration of such models into company workflows has been slow. This is due to a number of reasons.

  1. Being risk averse due to the giddying pace of development. Model architectures, model artifacts (e.g, different flavours of GPT models), and tools built atop these approaches is a very dynamic space, with reports of companies often evaluating build-vs-buy decisions on a weekly basis. Many companies are waiting for waters to still, and stable models to emerge before adoption.

  2. Being risk averse to non-deterministic nature and lack of interpretability of generative AI (e.g. in healthcare with strict regulations and costs of mis-predicting)

  3. Lack of culture/expertise in being able to run dynamic experiments at scale. Many industries/management decisions are top-down, and not conducive to bottoms-up approach. Thus experimentation is not always encouraged/financed, which makes it harder to nurture inhouse experience for productionisng LLMs.

  4. Hardware requirements are daunting, and often unaffordable, often requiring multiple GPUs operated in a distributed manner. This is formidably expensive for many, and is difficult to get budget approval of that kind of hardware requirements in experimentation phases of projects. (Note there are several third party efforts focused on simplifying distributed training/inference)

  5. Within a single company, LLM applications are developed in silos, making reuse and maintenance hard. Companies that invest in managed ML services/platforms are able to produce off the shelf ML components. In the absence of such centralised efforts, it is challenging for individual teams to secure the correct resources/expertise to quickly & reliably apply LLMs.

Mandatory disclaimer: I do see the winds of change, and believe the adoption will improve soon.

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 16: Appreciation post for a running companion app I discovered
Payal Mitra

If you wish to start, or improve your runs - regardless of what stage you are at, I would recommend trying the Run with Nike app. Up until a month ago, my longest runs would be 5km. In the last 4 weeks, I went from running 5km to 13km (runner's high is real!).

I've enjoyed my runs for a while, and did most of the training on my own without any app (but Spotify). I do recognise my body's improved athletic ability was due to a combination of better consistency (in exercise, sleep, meals), and a diversification of workouts (strength training, yoga, cardio, etc).

However, my last two long runs have been far more effortless and enjoyable than I would have thought possible. Coach Bennet (Nike Running Global Head Coach) from Run with Nike app has been transformational! Here's why I would recommend it:

  • Reason #1: Tailored to everyone on the running spectrum - whether one is starting with their first ever run, or one is a marathon runner. No run/goal is too small or big.

  • Reason #2: The app groups the guided runs in different categories that do a great job at covering all the reasons someone might want to run a particular run. E.g. Distance, time based runs, strategies, but also mindful runs, play, recovery etc.

  • Reason #3: Helps one keep pace properly and progressively during the run, so that one does not tire out. I have been finishing my runs more energetic than before.

  • Reason #4: Coach Bennet's prerecorded voice track is synced to your run's distance. So it really does feel like personalised coaching, while also leaving precious gaps for silence in between.

  • Reason #5: Focuses on mental gymnastics in the runner's mind to make run easier.

  • Reason #6: Syncs (most likely) with your music audio app

  • Reason #7: Get in touch with yourself. It is a privilege to have almost personalised coaching and some silence, that helps the runner listen & converse with oneself.

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 15: Reflections from when managing others did not go my way
Payal Mitra

As is law, we learn from mistakes and reflections. Here are three reflections from managing 'others'. I talk about managing my manager, peers, and junior colleagues.

I took away some important lessons about people management, and organisation of interactions that can drive outcomes from the times where I did not know better.

i. Managing your manager - Regular pulse checks and align to business objective

Communicate progress updates in a timely manner. Pulse-check how the work is perceived and if it is aligned to overall business objectives. Do not assume efforts are implicitly understood, and make sure to extract feedback. I found this to be particularly relevant when managing oneself in research/independent environments, to course correct or communicate when needed. I waited too long this one time, and it resulted in an average grade. I felt it was unfair, but in hindsight I know what I can improve now.

ii. Managing peers (guilds/working group) - keep short timelines, bound agenda

While leading guilds in the workplace, I found that facilitating a bunch of self-motivated folks, especially in volunteer efforts could be harder when the agenda of the working group is not clear, and if the timelines are too long. In the end, the group achieved good outcomes, but we had periods of low energy and I felt burnt out trying to deliver.

iii. Managing junior colleagues - encourage ownership + giving regular updates

Of course you want them to feel inspired and learn, but you don't want to over-manage. Have a learning log/plan that you both regularly check in on - this can be flexible, but it's very helpful to write intentions, progress and diversions down. Give ownership of small independent tasks within larger projects. The extra accountability+responsibility accelerates growth

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 14: Observations of C-suite behaviours that inspire outcomes - that you can apply too
Payal Mitra

These are behaviour patterns that signal effective leadership, at all rungs. They inspire trust in management and drive successful outcomes. This is from observation and applied experience (I'm no CxO, but I take ownership). #everyoneIsALeaderClub

1.Regular Communications (Thoughtful, informative, and concise)

Warren Buffet is known for his excellent shareholder letters at Berkshire Hathway. Alfred Sloan of GM was well known for thoughtful letters post management meetings.

I have observed tangible energy transfers when this is done consistently and thoughtfully at my workplace by others or me, at any level. Make it a fortnightly, monthly letter/postcard to all. Should include both good and bad (and actions).

2.Strong people hiring skills

This is often touted as the one of the most important roles of CxOs, founders, etc, by the leaders themselves. A behaviour that should propagate all along the chain.

3.Listening to internal grievances, and then acting

I participated in the 2nd set of internal listening tours organised by the CTO (of a large firm). I liked that this was not a one-off gesture sans accountability. I could clearly see how the top 1-2 themes from the previous year's tours were actually implemented.

4.Everyone is an owner

Everyone under/alongside you in the management chain is a leader/entrepreneur/change-driver, etc. They are responsible and accountable.

5.Purposeful, uniting 'mission' statements for big undertakings

A good inheritable tagline underlying and steering purpose is very underrated. It shows profound understanding of the why and an ability to breakdown grand visions into lowest common divisor tasks.

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 13: Recipe I would use if I were starting my Deep Learning (for natural language) journey today
Payal Mitra

Don't let the dizzying pace of advancements in the ChatGPTera overwhelm. Be inspired by it, accelerate your learning & build cool stuff from shoulders of communities.

  1. Youtube lectures. Find a youtuber (or a blogger) you enjoy learning from. There are great content-creators for almost everything. However, as this is my advice to my alternate-universe-self, I would strongly urge Andrej Karpathy's channel for what I think provides super cool hands-on learning by pulling (and reconstructing) ML stuff apart, decorated with wonder, and grounded with some math intuition.

  2. Build! This is non-negotiable. At least, tweak tutorials towards applications that interest you. Collaborate and build with others, makes it fun & purposeful!

  3. Follow your wonder. I recently stumbled upon 3Blue1Brown's channel. What a treat! I revisited information theory with wordle

  4. Keep your intuition of math and data strong, especially since this speaks to you.

  5. Wisely choose a structured online course with assignments if you have something very particular to learn - new frameworks, general foundational math/applications for neural networks (deeplearning.ai), niche use-cases (Biomedical AI), etc

  6. Github - Let your code live (and die) in public (I still largely fail)

  7. Write - personal notes. Document your learning for yourself

  8. Write - publicly, online. Stackoverflow, possibly blogs, etc. (Let it be useful. Don't parakeet a tutorial or blog piece. Write to capture your aha moments, tips, or to explain/teach concepts, or to share a cool application you made)

Whilst I'd want my alternate self to learn the modern techniques that are state of the art for a reason, I would also insist on strong foundations and first principles learning with a healthy balance of math+building. By means of a bad analogy - the study of concepts behind the first electric motors did not cease to be relevant to an aspiring electrical engineer today. To be fair, I am biased due to my academic training, where physics concepts were timeless. On an extreme, I even vouch for the lab sessions where we'd would program 8085 microprocessors and CROs from 1980s.

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 12: Game-changing tools to elevate and set your Data Science or ML project for success
Payal Mitra

One grossly under-emphasised skill for ML projects for both the Data Scientist and the team, is quick, interactive prototyping of an idea.

Five years ago building web services and apps for solely demo purposes of ML tools was painful and less-accessible, often requiring more unicorn data scientists with hack-y full-stack skills. This has dramatically changed in the past couple of years with beautiful efforts and frameworks to help build quick web apis for your ML models (e.g. FastAPI), and also interactive web app prototypes for your ML model or pipeline (e.g.: Streamlit, Gradio) in next to no time.

Why is this important?

  • setting projects up for success with better communication of ideas to stakeholders

  • helps in extracting feedback early in project from stakeholders (decision makers + users/customers)

  • building end to end prototypes tests ideas and holes, and can make iterations stronger and faster

  • better communication b/w cross-disciplinary team members - UX, data scientists. engineers, product managers

  • can reduce time to production greatly

    I have lost count of the number of times a certain ML project either does not move past the first experimental models, or moves too slowly. Much too often we only see lifeless tables of precision recall values, or slideware telling the story of the Data Science endeavour/intervention. Undoubtedly these tables store the most value and are the output of great skill+time to build and test ML models. However, would it not be transformative if we could bundle that sentiment classification model as a web service and expose it over an api to demonstrate its value to stakeholders? Or if we could have prospective users clicking on the cool Generative AI (ChatGPT-esque) tool to get feedback from users? I have found prototyping tools to be the most amazing research aids to ensure that what we build is what users need and want. It inspires the team and keeps the idea flywheel whirring.

My two favourite tools here would be Streamlit, and FastAPI - and I think they are a good place to start, with active communities and great tutorials.

This is not an exhaustive list, and there are many great tools out there for you to try. Essentially, search for tools and frameworks that help in building web apis and apps for demoing and testing ML capabilities.

Try it (if you haven't already), and you will see how being able to quickly deploy interactive apps and services that bring your demo Machine Learning capability to life is a real game changer.

0

Atomic Essay

Payal Mitra
Payal Mitra
Seeking and Sharing | AI Practitioner | Exploring Writing to test my curiosities
1y ago
Day 11: AI taking over human jobs? Not yet.
Payal Mitra

I'm adding to the text jungle on the 'AI taking over jobs' debate, because my mulling continues. Here are two aspects that I find myself filing my musings under:

Where do I stand in the AI threatens our jobs Q? I am on the optimistic side.
Where do I think future opportunity/demand for skills lie? Given space constraints on this atomic essay, reserving this for a future piece. However, my underlying beliefs are:

  • The value of human ingenuity, knowledge, empathy will not turn us redundant.

  • AI will change the nature of work & how we push ourselves to innovate, differentiate

  • I hope it pushes us to re-evaluate what we consider important, & how to reprioritise certain aspects of nature & human life that are devalued in today's constant hyper-consumerism, hyperfinancialisation and chase for higher economic growth.

0

Atomic Essay