The best Large Language Models of September 2024

The TIMETOACT GROUP LLM Benchmarks highlight the most powerful AI language models for digital product development. Discover which large language models performed best in September.

LLM Benchmarks | September 2024

The benchmarks evaluate the models in terms of their suitability for digital product development. The higher the score, the better.

☁️ - Cloud models with proprietary license
✅ - Open source models that can be run locally without restrictions
🦙 - Local models with Llama license

ModelCodeCrmDocsIntegrateMarketingReasonFinalCostSpeed
GPT o1-preview v1/2024-09-12 ☁️9592949688879252.32 €0.08 rps
GPT o1-mini v1/2024-09-12 ☁️939694858287908.15 €0.16 rps
Google Gemini 1.5 Pro v2 ☁️8697941007874881.00 €1.18 rps
GPT-4o v1/2024-05-13 ☁️9096100897874881.21 €1.44 rps
GPT-4o v3/dyn-2024-08-13 ☁️9097100817978881.22 €1.21 rps
GPT-4 Turbo v5/2024-04-09 ☁️8699981008843862.45 €0.84 rps
GPT-4o v2/2024-08-06 ☁️908497928259840.63 €1.49 rps
Google Gemini 1.5 Pro 0801 ☁️8492791007074830.90 €0.83 rps
Qwen 2.5 72B Instruct ⚠️7992941007159830.10 €0.66 rps
Llama 3.1 405B Hermes 3🦙6893891008853820.54 €0.49 rps
GPT-4 v1/0314 ☁️908898708845807.04 €1.31 rps
GPT-4 v2/0613 ☁️908395708845787.04 €2.16 rps
Claude 3 Opus ☁️6988100787658784.69 €0.41 rps
Claude 3.5 Sonnet ☁️728389858058780.94 €0.09 rps
GPT-4 Turbo v4/0125-preview ☁️6697100857543782.45 €0.84 rps
GPT-4o Mini ☁️6387807010065780.04 €1.46 rps
Meta Llama3.1 405B Instruct🦙819392707548762.39 €1.16 rps
GPT-4 Turbo v3/1106-preview ☁️667598708860762.46 €0.68 rps
DeepSeek v2.5 236B ⚠️578091788857750.03 €0.42 rps
Google Gemini 1.5 Flash v2 ☁️649689758144750.06 €2.01 rps
Google Gemini 1.5 Pro 0409 ☁️689796857526740.95 €0.59 rps
Meta Llama 3.1 70B Instruct f16🦙748990707548741.79 €0.90 rps
GPT-3.5 v2/0613 ☁️688173818150720.34 €1.46 rps
Meta Llama 3 70B Instruct🦙818384608145720.06 €0.85 rps
Mistral Large 123B v2/2407 ☁️687968757570720.86 €1.02 rps
Google Gemini 1.5 Pro 0514 ☁️7396791002560721.07 €0.92 rps
Google Gemini 1.5 Flash 0514 ☁️3297100757252710.06 €1.77 rps
Google Gemini 1.0 Pro ☁️668683788828710.37 €1.36 rps
Meta Llama 3.2 90B Vision🦙748487787132710.23 €1.10 rps
GPT-3.5 v3/1106 ☁️687071787858700.24 €2.33 rps
GPT-3.5 v4/0125 ☁️638771787843700.12 €1.43 rps
Qwen1.5 32B Chat f16 ⚠️709082787820690.97 €1.66 rps
Cohere Command R+ ☁️638076707058690.83 €1.90 rps
Gemma 2 27B IT ⚠️617287708932690.07 €0.90 rps
Mistral 7B OpenChat-3.5 v3 0106 f16 ✅688767708825670.32 €3.39 rps
Gemma 7B OpenChat-3.5 v3 0106 f16 ✅636784608146670.21 €5.09 rps
Meta Llama 3 8B Instruct f16🦙796268708041670.32 €3.33 rps
Mistral 7B OpenChat-3.5 v2 1210 f16 ✅637372698830660.32 €3.40 rps
Mistral 7B OpenChat-3.5 v1 f16 ✅587272708833650.49 €2.20 rps
GPT-3.5-instruct 0914 ☁️479269628833650.35 €2.15 rps
GPT-3.5 v1/0301 ☁️558269788226650.35 €4.12 rps
Llama 3 8B OpenChat-3.6 20240522 f16 ✅765176608838650.28 €3.79 rps
Mistral Nemo 12B v1/2407 ☁️5458511007549640.03 €1.22 rps
Meta Llama 3.2 11B Vision🦙707165707136640.04 €1.49 rps
Starling 7B-alpha f16 ⚠️586667708834640.58 €1.85 rps
Llama 3 8B Hermes 2 Theta🦙617374708516630.05 €0.55 rps
Yi 1.5 34B Chat f16 ⚠️477870708626631.18 €1.37 rps
Claude 3 Haiku ☁️646964707535630.08 €0.52 rps
Meta Llama 3.1 8B Instruct f16🦙577462707432610.45 €2.41 rps
Qwen2 7B Instruct f32 ⚠️508181606631610.46 €2.36 rps
Mistral Small v3/2409 ☁️437571757526610.06 €0.81 rps
Claude 3 Sonnet ☁️724174707828610.95 €0.85 rps
Mixtral 8x22B API (Instruct) ☁️536262100757600.17 €3.12 rps
Mistral Pixtral 12B ✅536973606440600.03 €0.83 rps
Codestral Mamba 7B v1 ✅5366511007117600.30 €2.82 rps
Anthropic Claude Instant v1.2 ☁️587565756516592.10 €1.49 rps
Cohere Command R ☁️456657708427580.13 €2.50 rps
Anthropic Claude v2.0 ☁️635255608434582.19 €0.40 rps
Qwen1.5 7B Chat f16 ⚠️568160506036570.29 €3.76 rps
Mistral Large v1/2402 ☁️374970788425570.58 €2.11 rps
Microsoft WizardLM 2 8x22B ⚠️487679506222560.13 €0.70 rps
Qwen1.5 14B Chat f16 ⚠️505851708422560.36 €3.03 rps
Anthropic Claude v2.1 ☁️295859787532552.25 €0.35 rps
Llama2 13B Vicuna-1.5 f16🦙503755608237530.99 €1.09 rps
Mistral 7B Instruct v0.1 f16 ☁️347169596223530.75 €1.43 rps
Mistral 7B OpenOrca f16 ☁️545776257827530.41 €2.65 rps
Meta Llama 3.2 3B🦙527166704414530.01 €1.25 rps
Google Recurrent Gemma 9B IT f16 ⚠️582771605623490.89 €1.21 rps
Codestral 22B v1 ✅384744786613480.06 €4.03 rps
Llama2 13B Hermes f16🦙502437746042481.00 €1.07 rps
IBM Granite 34B Code Instruct f16 ☁️63493470577471.07 €1.51 rps
Mistral Small v2/2402 ☁️33424592568460.06 €3.21 rps
DBRX 132B Instruct ⚠️433943775910450.26 €1.31 rps
Mistral Medium v1/2312 ☁️414344616212440.81 €0.35 rps
Meta Llama 3.2 1B🦙324033406851440.02 €1.69 rps
Llama2 13B Puffin f16🦙371544705639434.70 €0.23 rps
Mistral Small v1/2312 (Mixtral) ☁️10676352568430.06 €2.21 rps
Microsoft WizardLM 2 7B ⚠️533442595313420.02 €0.89 rps
Mistral Tiny v1/2312 (7B Instruct v0.2) ☁️22475938628390.05 €2.39 rps
Gemma 2 9B IT ⚠️452547346813380.02 €0.88 rps
Meta Llama2 13B chat f16🦙22381760756360.75 €1.44 rps
Mistral 7B Zephyr-β f16 ✅37344659294350.46 €2.34 rps
Meta Llama2 7B chat f16🦙223320605018340.56 €1.93 rps
Mistral 7B Notus-v1 f16 ⚠️10542552484320.75 €1.43 rps
Orca 2 13B f16 ⚠️182232226720300.95 €1.14 rps
Mistral 7B v0.1 f16 ☁️0948535212290.87 €1.23 rps
Mistral 7B Instruct v0.2 f16 ☁️11305412588290.96 €1.12 rps
Google Gemma 2B IT f16 ⚠️332816571520280.30 €3.54 rps
Microsoft Phi 3 Medium 4K Instruct 14B f16 ⚠️5343011478220.82 €1.32 rps
Orca 2 7B f16 ⚠️2202620524210.78 €1.38 rps
Google Gemma 7B IT f16 ⚠️0009620120.99 €1.08 rps
Meta Llama2 7B f16🦙05223282100.95 €1.13 rps
Yi 1.5 9B Chat f16 ⚠️042990881.41 €0.76 rps

Can the model generate code and help with programming?

The estimated cost of running the workload. For cloud-based models, we calculate the cost according to the pricing. For on-premises models, we estimate the cost based on GPU requirements for each model, GPU rental cost, model speed, and operational overhead.

How well does the model support work with product catalogs and marketplaces?

How well can the model work with large documents and knowledge bases?

Can the model easily interact with external APIs, services and plugins?

How well can the model support marketing activities, e.g. brainstorming, idea generation and text generation?

How well can the model reason and draw conclusions in a given context?

The "Speed" column indicates the estimated speed of the model in requests per second (without batching). The higher the speed, the better.

ChatGPT o1 models are the best

OpenAI has released a radically new type of the model called o1-preview that is followed by o1-mini. These unique models differ from all the other LLM models out there - they run their own chain of thought routine for each request. This allow the model to decompose complex problems in smaller tasks and really think the answers through.

That approach, for example, shines in complex full-stack software engineering challenges. o1, if compared to the “ordinary” GPT-4 feels like an experienced Middle Level Software Engineer that requires surprisingly little hand-holding.

There is one downside in this “chain of thought under the hood” process. O1 produces high quality results, but these results take time and cost a lot more. Just look at the comparative pricing within the Cost column.

We are looking forward to see other LLM vendors take a note of this trick and release their own versions of LLMs with tuned chain-of-thought routine.

Google Gemini 1.5 Pro v 002 - TOP 3

While speaking of the top results and cloud vendors, there is another new model in the TOP-3. Google has somehow managed to catch up with the rate of the progress and release highly competitive model - Gemini 1.5 Pro v 002.

This model systematically improves over the previous version in multiple categories: Code, CRM, Docs, and Marketing texts. It is also the cheapest model in the TOP-6 of our benchmark.

Practitioners already praise this model for great multi-lingual skills, while users of Google Cloud are happy to have top-tier LLM available in their cloud.

For a long time, it felt like OpenAI and Anthropic are the only companies that can really push the state of the art in top-tier LLM models. It also felt like large mammoth companies are just too slow and old-school to release something worthy. Google was eventually able to prove this wrong.

This is how the progress of Google models looks over the time:

Now it doesn’t feel out of the ordinary to expect models of similar quality from Amazon or Microsoft. Perhaps, this will spur forward a round of competition with further price drops and further quality improvements.
Enough with the cloud vendors. Let’s talk about local models now.

(Local models are the models that you can download and run on your own hardware.)

Qwen 2.5 and DeepSeek 2.5

Recently released Qwen 2.5 Instruct is surprisingly good. This is the first local model that beats Claude 3.5 Sonnet on our business tasks. It also costs less than the other LLM models in the top.

Starting from this benchmark we’ll use OpenRouter pricing as the base price for locally-capable LLM models. This allows to estimate workload costs based on the real-world market. It also factors in any meaningful performance optimisations that LLM vendors are willing to use to improve their margins.

Qwen 2.5 72B diligently follows instructions (if compared to Sonnet 3.5 or older GPT-4 versions) and has a decent Reason capability. This Chinese model has gaps in Code and Marketing capabilities.

DeepSeek 2.5 didn’t perform nearly as well in our product benchmarks, despite having a huge size of 236B parameters. It runs roughly on the level of older versions of GPT-4 and Gemini 1.5 Pro.

These actually are outstanding news: more and more local models reach the level of GPT-4 Turbo intelligence. And the fact that a smaller Qwen 72B model has beaten it by a big margin - is worth a separate celebration 🚀

We think, this is not the last celebration of this kind for this year.

Llama 3.2 - Mediocre results, but there is a small nuance

Meta has just released their new versions of Llama - 3.2 model range.

Larger models are now multi-modal. This happened at the cost of the cognitive capabilities in text-driven business tasks, if compared to the previous model versions. Llama 3.2 is still far from the top.

If we look at the table:

  • Llama 3.2 90B Vision works on the level of Llama 3/3.1 70B but with worse Reason.

  • Llama 3.2 11B Vision works on the level of previous 8B, but with worse reason.

This doesn’t make the new models worse - they have more capabilities now. Our benchmark currently tests only text-based business tasks. Vision tasks will be added in v2.

Having said that, there is a small nuance that really makes this Llama 3.2 release outstanding. Size of that nuance is 1B and 3B. These are the sizes of new tiny Llama 3.2 models that are designed to run in resource-constrained environments and on the edge (optimised for ARM processors, Qualcomm and MediaTek hardware). Despite resource constraints, these models feature 128k token context and surprisingly high response quality in business tasks.

For example, do you remember a huge DBRX 132B Instruct model that claimed to be “a new state-of-the-art for established open LLMs”? Well, Llama 3.2 1B model catches up with it in our benchmark and 3B beats it by a big margin. Just look at the neighbours of these models on this table:

Keep in mind that these benchmarks results are for the base Llama versions. Customised fine-tunes tend to improve overall scores even further.

As you can see, the progress doesn’t stand still. We’ll be waiting for the continuation of the trend where more and more companies manage to package better cognitive capability in smaller models.

To visualise such a trend, we’ve plotted all releases of locally-capable models over the timeline. Then we’ve grouped them together based on the rough hardware requirements for running them. For each group we’ve computed current trend (linregress)

Note: this grouping is very rough. We are using for most frequent hardware combinations that we’ve seen among our customers and within the AI Research. We are also assuming that we are running inference under fp16 without any further quantisations and with enough spare VRAM to keep some context around.

Here are a few observations.

  • All models are getting better over the time - both small and big ones.
  • Interesting large models started showing up on the radar only this year.
  • Large models currently have the fastest rage of improvement.

These observations are obvious. You don’t need a chart to figure them out. However visualisations make rate of the progress more comprehensible. It could then be translated to customers and accounted for in long-term plans.

Transform Your Digital Projects with the Best AI Language Models!

Discover the transformative power of the best LLMs and revolutionize your digital products with AI! Stay future-focused, boost efficiency, and gain a clear competitive edge. We help you elevate your business value to the next level.

 

* required

Wir verwenden die von Ihnen an uns gesendeten Angaben nur, um auf Ihren Wunsch hin mit Ihnen Kontakt im Zusammenhang mit Ihrer Anfrage aufzunehmen. Alle weiteren Informationen können Sie unseren Datenschutzhinweisen entnehmen.

Blog
Blog

ChatGPT & Co: LLM Benchmarks for October

Find out which large language models outperformed in the October 2024 benchmarks. Stay informed on the latest AI developments and performance metrics.

Blog
Blog

ChatGPT & Co: LLM Benchmarks for November

Find out which large language models outperformed in the November 2024 benchmarks. Stay informed on the latest AI developments and performance metrics.

Rinat AbdullinRinat AbdullinBlog
Blog

Open-sourcing 4 solutions from the Enterprise RAG Challenge

Our RAG competition is a friendly challenge different AI Assistants competed in answering questions based on the annual reports of public companies.

Rinat AbdullinRinat AbdullinBlog
Blog

LLM Performance Series: Batching

Beginning with the September Trustbit LLM Benchmarks, we are now giving particular focus to a range of enterprise workloads. These encompass the kinds of tasks associated with Large Language Models that are frequently encountered in the context of large-scale business digitalization.

Rinat AbdullinRinat AbdullinBlog
Blog

Strategic Impact of Large Language Models

This blog discusses the rapid advancements in large language models, particularly highlighting the impact of OpenAI's GPT models.

Jörg EgretzbergerJörg EgretzbergerBlog
Blog

8 tips for developing AI assistants

AI assistants for businesses are hype, and many teams were already eagerly and enthusiastically working on their implementation. Unfortunately, however, we have seen that many teams we have observed in Europe and the US have failed at the task. Read about our 8 most valuable tips, so that you will succeed.

Martin WarnungMartin WarnungBlog
Blog

Common Mistakes in the Development of AI Assistants

How fortunate that people make mistakes: because we can learn from them and improve. We have closely observed how companies around the world have implemented AI assistants in recent months and have, unfortunately, often seen them fail. We would like to share with you how these failures occurred and what can be learned from them for future projects: So that AI assistants can be implemented more successfully in the future!

Aqeel AlazreeBlog
Blog

Part 1: Data Analysis with ChatGPT

In this new blog series we will give you an overview of how to analyze and visualize data, create code manually and how to make ChatGPT work effectively. Part 1 deals with the following: In the data-driven era, businesses and organizations are constantly seeking ways to extract meaningful insights from their data. One powerful tool that can facilitate this process is ChatGPT, a state-of-the-art natural language processing model developed by OpenAI. In Part 1 pf this blog, we'll explore the proper usage of data analysis with ChatGPT and how it can help you make the most of your data.

Blog
Blog

Second Place - AIM Hackathon 2024: Trustpilot for ESG

The NightWalkers designed a scalable tool that assigns trustworthiness scores based on various types of greenwashing indicators, including unsupported claims and inaccurate data.

Felix KrauseBlog
Blog

AIM Hackathon 2024: Sustainability Meets LLMs

Focusing on impactful AI applications, participants addressed key issues like greenwashing detection, ESG report relevance mapping, and compliance with the European Green Deal.

Blog
Blog

SAM Wins First Prize at AIM Hackathon

The winning team of the AIM Hackathon, nexus. Group AI, developed SAM, an AI-powered ESG reporting platform designed to help companies streamline their sustainability compliance.

Blog
Blog

Third Place - AIM Hackathon 2024: The Venturers

ESG reports are often filled with vague statements, obscuring key facts investors need. This team created an AI prototype that analyzes these reports sentence-by-sentence, categorizing content to produce a "relevance map".

Matus ZilinskyBlog
Blog

Creating a Social Media Posts Generator Website with ChatGPT

Using the GPT-3-turbo and DALL-E models in Node.js to create a social post generator for a fictional product can be really helpful. The author uses ChatGPT to create an API that utilizes the openai library for Node.js., a Vue component with an input for the title and message of the post. This article provides step-by-step instructions for setting up the project and includes links to the code repository.

Rinat AbdullinRinat AbdullinBlog
Blog

The Intersection of AI and Voice Manipulation

The advent of Artificial Intelligence (AI) in text-to-speech (TTS) technologies has revolutionized the way we interact with written content. Natural Readers, standing at the forefront of this innovation, offers a comprehensive suite of features designed to cater to a broad spectrum of needs, from personal leisure to educational support and commercial use. As we delve into the capabilities of Natural Readers, it's crucial to explore both the advantages it brings to the table and the ethical considerations surrounding voice manipulation in TTS technologies.

Aqeel AlazreeBlog
Blog

Part 4: Save Time and Analyze the Database File

ChatGPT-4 enables you to analyze database contents with just two simple steps (copy and paste), facilitating well-informed decision-making.

Aqeel AlazreeBlog
Blog

Part 3: How to Analyze a Database File with GPT-3.5

In this blog, we'll explore the proper usage of data analysis with ChatGPT and how you can analyze and visualize data from a SQLite database to help you make the most of your data.

Rinat AbdullinRinat AbdullinBlog
Blog

Let's build an Enterprise AI Assistant

In the previous blog post we have talked about basic principles of building AI assistants. Let’s take them for a spin with a product case that we’ve worked on: using AI to support enterprise sales pipelines.

Rinat AbdullinRinat AbdullinBlog
Blog

So You are Building an AI Assistant?

So you are building an AI assistant for the business? This is a popular topic in the companies these days. Everybody seems to be doing that. While running AI Research in the last months, I have discovered that many companies in the USA and Europe are building some sort of AI assistant these days, mostly around enterprise workflow automation and knowledge bases. There are common patterns in how such projects work most of the time. So let me tell you a story...

Workshop
Workshop

AI Workshops for Companies

Whether it's the basics of AI, prompt engineering, or potential scouting: our diverse AI workshop offerings provide the right content for every need.

TIMETOACT
Martin LangeMartin LangeBlog
Checkliste als Symbol für die verschiedenen To Dos im Bereich Lizenzmanagement
Blog

License Management – Everything you need to know

License management is not only relevant in terms of compliance but can also minimize costs and risks. Read more in the article.

Felix KrauseBlog
Blog

License Plate Detection for Precise Car Distance Estimation

When it comes to advanced driver-assistance systems or self-driving cars, one needs to find a way of estimating the distance to other vehicles on the road.

Aqeel AlazreeBlog
Blog

Database Analysis Report

This report comprehensively analyzes the auto parts sales database. The primary focus is understanding sales trends, identifying high-performing products, Analyzing the most profitable products for the upcoming quarter, and evaluating inventory management efficiency.

Rinat AbdullinRinat AbdullinBlog
Blog

5 Inconvenient Questions when hiring an AI company

This article discusses five questions you should ask when buying an AI. These questions are inconvenient for providers of AI products, but they are necessary to ensure that you are getting the best product for your needs. The article also discusses the importance of testing the AI system on your own data to see how it performs.

TIMETOACT
Referenz
Referenz

Managed service support for optimal license management

To ensure software compliance, TIMETOACT supports FUNKE Mediengruppe with a SAM Managed Service for Microsoft, Adobe, Oracle and IBM.

TIMETOACT
Referenz
Referenz

Interactive online portal identifies suitable employees

TIMETOACT digitizes several test procedures for KI.TEST to determine professional intelligence and personality.

Branche
Branche

Artificial Intelligence in Treasury Management

Optimize treasury processes with AI: automated reports, forecasts, and risk management.

TIMETOACT
Referenz
Referenz

Standardized data management creates basis for reporting

TIMETOACT implements a higher-level data model in a data warehouse for TRUMPF Photonic Components and provides the necessary data integration connection with Talend. With this standardized data management, TRUMPF will receive reports based on reliable data in the future and can also transfer the model to other departments.

Daniel WellerBlog
Blog

Revolutionizing the Logistics Industry

As the logistics industry becomes increasingly complex, businesses need innovative solutions to manage the challenges of supply chain management, trucking, and delivery. With competitors investing in cutting-edge research and development, it is vital for companies to stay ahead of the curve and embrace the latest technologies to remain competitive. That is why we introduce the TIMETOACT Logistics Simulator Framework, a revolutionary tool for creating a digital twin of your logistics operation.

TIMETOACT
Referenz
Referenz

Flexibility in the data evaluation of a theme park

With the support of TIMETOACT, an theme park in Germany has been using TM1 for many years in different areas of the company to carry out reporting, analysis and planning processes easily and flexibly.

Referenz
Referenz

Automated Planning of Transport Routes

Efficient transport route planning through automation and seamless integration.

Rinat AbdullinRinat AbdullinBlog
Blog

Using NLP libraries for post-processing

Learn how to analyse sticky notes in miro from event stormings and how this analysis can be carried out with the help of the spaCy library.

Laura GaetanoBlog
Blog

Using a Skill/Will matrix for personal career development

Discover how a Skill/Will Matrix helps employees identify strengths and areas for growth, boosting personal and professional development.

Sebastian BelczykBlog
Blog

Building A Shell Application for Micro Frontends | Part 4

We already have a design system, several micro frontends consuming this design system, and now we need a shell application that imports micro frontends and displays them.

Bernhard SchauerBlog
Blog

ADRs as a Tool to Build Empowered Teams

Learn how we use Architecture Decision Records (ADRs) to build empowered, autonomous teams, enhancing decision-making and collaboration.

Peter SzarvasPeter SzarvasBlog
Blog

Why Was Our Project Successful: Coincidence or Blueprint?

“The project exceeded all expectations,” is one among our favourite samples of the very positive feedback from our client. Here's how we did it!

Christian FolieBlog
Blog

The Power of Event Sourcing

This is how we used Event Sourcing to maintain flexibility, handle changes, and ensure efficient error resolution in application development.

Jonathan ChannonBlog
Blog

Tracing IO in .NET Core

Learn how we leverage OpenTelemetry for efficient tracing of IO operations in .NET Core applications, enhancing performance and monitoring.

Rinat AbdullinRinat AbdullinBlog
Blog

Consistency and Aggregates in Event Sourcing

Learn how we ensures data consistency in event sourcing with effective use of aggregates, enhancing system reliability and performance.

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 10

Discover Agents and Mailboxes in F#. Build responsive applications using these powerful concurrency tools in functional programming.

Rinat AbdullinRinat AbdullinBlog
Blog

Process Pipelines

Discover how process pipelines break down complex tasks into manageable steps, optimizing workflows and improving efficiency using Kanban boards.

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 11

Learn type inference and generic functions in F#. Boost efficiency and flexibility in your code with these essential programming concepts.

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 12

Explore reflection and meta-programming in F#. Learn how to dynamically manipulate code and enhance flexibility with advanced techniques.

Rinat AbdullinRinat AbdullinBlog
Blog

Innovation Incubator at TIMETOACT GROUP Austria

Discover how our Innovation Incubator empowers teams to innovate with collaborative, week-long experiments, driving company-wide creativity and progress.

Rinat AbdullinRinat AbdullinBlog
Blog

Innovation Incubator Round 1

Team experiments with new technologies and collaborative problem-solving: This was our first round of the Innovation Incubator.

Rinat AbdullinRinat AbdullinBlog
Blog

Announcing Domain-Driven Design Exercises

Interested in Domain Driven Design? Then this DDD exercise is perfect for you!

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 8

Discover Units of Measure and Type Providers in F#. Enhance data management and type safety in your applications with these powerful tools.

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 9

Explore Active Patterns and Computation Expressions in F#. Enhance code clarity and functionality with these advanced techniques.

Blog
Blog

My Workflows During the Quarantine

The current situation has deeply affected our daily lives. However, in retrospect, it had a surprisingly small impact on how we get work done at TIMETOACT GROUP Austria.

Rinat AbdullinRinat AbdullinBlog
Blog

Learning + Sharing at TIMETOACT GROUP Austria

Discover how we fosters continuous learning and sharing among employees, encouraging growth and collaboration through dedicated time for skill development.

Laura GaetanoBlog
Blog

My Weekly Shutdown Routine

Discover my weekly shutdown routine to enhance productivity and start each week fresh. Learn effective techniques for reflection and organization.

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 7

Explore LINQ and query expressions in F#. Simplify data manipulation and enhance your functional programming skills with this guide.

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 3

Dive into F# data structures and pattern matching. Simplify code and enhance functionality with these powerful features.

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 4

Unlock F# collections and pipelines. Manage data efficiently and streamline your functional programming workflow with these powerful tools.

Sebastian BelczykBlog
Blog

Building a micro frontend consuming a design system | Part 3

In this blopgpost, you will learn how to create a react application that consumes a design system.

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 5

Master F# asynchronous workflows and parallelism. Enhance application performance with advanced functional programming techniques.

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 6

Learn error handling in F# with option types. Improve code reliability using F#'s powerful error-handling techniques.

Ian RussellIan RussellBlog
Blog

So, I wrote a book

Join me as I share the story of writing a book on F#. Discover the challenges, insights, and triumphs along the way.

Nina DemuthBlog
Blog

7 Positive effects of visualizing the interests of your team

Interests maps unleash hidden potentials and interests, but they also make it clear which topics are not of interest to your colleagues.

Daniel PuchnerBlog
Blog

How to gather data from Miro

Learn how to gather data from Miro boards with this step-by-step guide. Streamline your data collection for deeper insights.

Christian FolieBlog
Blog

Designing and Running a Workshop series: An outline

Learn how to design and execute impactful workshops. Discover tips, strategies, and a step-by-step outline for a successful workshop series.

Christian FolieBlog
Blog

Designing and Running a Workshop series: The board

In this part, we discuss the basic design of the Miro board, which will aid in conducting the workshops.

Felix KrauseBlog
Blog

Creating a Cross-Domain Capable ML Pipeline

As classifying images into categories is a ubiquitous task occurring in various domains, a need for a machine learning pipeline which can accommodate for new categories is easy to justify. In particular, common general requirements are to filter out low-quality (blurred, low contrast etc.) images, and to speed up the learning of new categories if image quality is sufficient. In this blog post we compare several image classification models from the transfer learning perspective.

Rinat AbdullinRinat AbdullinBlog
Blog

State of Fast Feedback in Data Science Projects

DSML projects can be quite different from the software projects: a lot of R&D in a rapidly evolving landscape, working with data, distributions and probabilities instead of code. However, there is one thing in common: iterative development process matters a lot.

Felix KrauseBlog
Blog

Part 2: Detecting Truck Parking Lots on Satellite Images

In the previous blog post, we created an already pretty powerful image segmentation model in order to detect the shape of truck parking lots on satellite images. However, we will now try to run the code on new hardware and get even better as well as more robust results.

Felix KrauseBlog
Blog

Part 1: Detecting Truck Parking Lots on Satellite Images

Real-time truck tracking is crucial in logistics: to enable accurate planning and provide reliable estimation of delivery times, operators build detailed profiles of loading stations, providing expected durations of truck loading and unloading, as well as resting times. Yet, how to derive an exact truck status based on mere GPS signals?

Laura GaetanoBlog
Blog

5 lessons from running a (remote) design systems book club

Last year I gifted a design systems book I had been reading to a friend and she suggested starting a mini book club so that she’d have some accountability to finish reading the book. I took her up on the offer and so in late spring, our design systems book club was born. But how can you make the meetings fun and engaging even though you're physically separated? Here are a couple of things I learned from running my very first remote book club with my friend!

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F# – Part 2

Explore functions, types, and modules in F#. Enhance your skills with practical examples and insights in this detailed guide.

Sebastian BelczykBlog
Blog

Composite UI with Design System and Micro Frontends

Discover how to create scalable composite UIs using design systems and micro-frontends. Enhance consistency and agility in your development process.

Sebastian BelczykBlog
Blog

Building and Publishing Design Systems | Part 2

Learn how to build and publish design systems effectively. Discover best practices for creating reusable components and enhancing UI consistency.

Daniel PuchnerBlog
Blog

Make Your Value Stream Visible Through Structured Logging

Boost your value stream visibility with structured logging. Improve traceability and streamline processes in your software development lifecycle.

Aqeel AlazreeBlog
Blog

Part 2: Data Analysis with powerful Python

Analyzing and visualizing data from a SQLite database in Python can be a powerful way to gain insights and present your findings. In Part 2 of this blog series, we will walk you through the steps to retrieve data from a SQLite database file named gold.db and display it in the form of a chart using Python. We'll use some essential tools and libraries for this task.

Daniel PuchnerBlog
Blog

How we discover and organise domains in an existing product

Software companies and consultants like to flex their Domain Driven Design (DDD) muscles by throwing around terms like Domain, Subdomain and Bounded Context. But what lies behind these buzzwords, and how these apply to customers' diverse environments and needs, are often not as clear. As it turns out it takes a collaborative effort between stakeholders and development team(s) over a longer period of time on a regular basis to get them right.

Ian RussellIan RussellBlog
Blog

Introduction to Functional Programming in F#

Dive into functional programming with F# in our introductory series. Learn how to solve real business problems using F#'s functional programming features. This first part covers setting up your environment, basic F# syntax, and implementing a simple use case. Perfect for developers looking to enhance their skills in functional programming.

Rinat AbdullinRinat AbdullinBlog
Blog

Machine Learning Pipelines

In this first part, we explain the basics of machine learning pipelines and showcase what they could look like in simple form. Learn about the differences between software development and machine learning as well as which common problems you can tackle with them.

Rinat AbdullinRinat AbdullinBlog
Blog

Event Sourcing with Apache Kafka

For a long time, there was a consensus that Kafka and Event Sourcing are not compatible with each other. So it might look like there is no way of working with Event Sourcing. But there is if certain requirements are met.

Christian FolieBlog
Blog

Running Hybrid Workshops

When modernizing or building systems, one major challenge is finding out what to build. In Pre-Covid times on-site workshops were a main source to get an idea about ‘the right thing’. But during Covid everybody got used to working remotely, so now the question can be raised: Is it still worth having on-site, physical workshops?

Felix KrauseBlog
Blog

Boosting speed of scikit-learn regression algorithms

The purpose of this blog post is to investigate the performance and prediction speed behavior of popular regression algorithms, i.e. models that predict numerical values based on a set of input variables.

Chrystal LantnikBlog
Blog

CSS :has() & Responsive Design

In my journey to tackle a responsive layout problem, I stumbled upon the remarkable benefits of the :has() pseudo-class. Initially, I attempted various other methods to resolve the issue, but ultimately, embracing the power of :has() proved to be the optimal solution. This blog explores my experience and highlights the advantages of utilizing the :has() pseudo-class in achieving flexible layouts.

Rinat AbdullinRinat AbdullinBlog
Blog

Inbox helps to clear the mind

I hate distractions. They can easily ruin my day when I'm in the middle of working on a cool project. They do that by overloading my mind, buzzing around inside me, and just making me tired. Even though we can think about several things at once, we can only do one thing at a time.

Rinat AbdullinRinat AbdullinBlog
Blog

Celebrating achievements

Our active memory can be like a cache of recently used data; fresh ideas & frustrations supersede older ones. That's why celebrating achievements is key for your success.

Ian RussellIan RussellBlog
Blog

Introduction to Web Programming in F# with Giraffe – Part 1

In this series we are investigating web programming with Giraffe and the Giraffe View Engine plus a few other useful F# libraries.

Ian RussellIan RussellBlog
Blog

Introduction to Web Programming in F# with Giraffe – Part 2

In this series we are investigating web programming with Giraffe and the Giraffe View Engine plus a few other useful F# libraries.

Ian RussellIan RussellBlog
Blog

Introduction to Web Programming in F# with Giraffe – Part 3

In this series we are investigating web programming with Giraffe and the Giraffe View Engine plus a few other useful F# libraries.

Balazs MolnarBalazs MolnarBlog
Blog

Learn & Share video Obsidian

Knowledge is very powerful. So, finding the right tool to help you gather, structure and access information anywhere and anytime, is rather a necessity than an option. You want to accomplish your tasks better? You want a reliable tool which is easy to use, extendable and adaptable to your personal needs? Today I would like to introduce you to the knowledge management system of my choice: Obsidian.