Entries by admin

MuFo: Branding and Visual Identity for Photography Museum

Original Source: https://abduzeedo.com/mufo-branding-and-visual-identity-photography-museum

MuFo: Branding and Visual Identity for Photography Museum
MuFo: Branding and Visual Identity for Photography Museum

abduzeedo0129—24

Explore the innovative branding and visual identity of Krakow’s MuFo, the only museum in Poland dedicated to photography. Uncover how MuFo’s unique design sets it apart.

In the heart of Krakow, the Museum of Photography, now rebranded as MuFo, stands as a beacon of cultural and visual innovation. This transformation is not just a change of name, but a complete overhaul of its visual identity, symbolizing the museum’s new direction and broader offerings.

MuFo, derived from the words Museum and Fotografia (Photography in Polish), exemplifies simplicity and memorability in branding. This concise, easily pronounceable name transcends language barriers, positioning the museum as an accessible cultural hub, not just in Poland but internationally.

The museum’s logo is a testament to thoughtful design. Focusing on universal concepts like vision, observation, and the photographic lens, it converges into a symbol resembling an eye. This emblem is not merely a static image; it represents the dynamic nature of photography and the museum’s role as a site of observation and discovery.

But MuFo’s visual identity extends beyond its logo. It is a comprehensive expression of the museum’s character. The identity is versatile, with variable layouts and forms that adapt to different contexts. This flexibility is crucial in a world where digital and urban spaces are saturated with information and imagery. MuFo’s use of a vivid color palette further enhances its visibility, ensuring that its communication is clear and impactful in a diverse urban landscape.

MuFo is more than a museum; it’s a cultural epicenter, a hub for exchanging ideas, and a showcase of rich photographic collections. Its design and branding reflect this status, making it a standout entity in Poland and a significant landmark on the cultural map of Europe.

The redesign of MuFo is a clear indication of the museum embracing the 21st century, where design goes hand in hand with functionality and cultural significance. It’s a place that captures the eye and the imagination at first glance, inviting visitors to explore the depths of photography and its impact on society. This project is a striking example of how visual identity can revitalize an institution, making it a beacon of cultural and artistic innovation.

Branding and visual identity artifacts

Artifact from the MuFo: Branding and Visual Identity for Photography Museum article on Abduzeedo

Artifact from the MuFo: Branding and Visual Identity for Photography Museum article on AbduzeedoArtifact from the MuFo: Branding and Visual Identity for Photography Museum article on Abduzeedo

Artifact from the MuFo: Branding and Visual Identity for Photography Museum article on Abduzeedo

Artifact from the MuFo: Branding and Visual Identity for Photography Museum article on AbduzeedoArtifact from the MuFo: Branding and Visual Identity for Photography Museum article on Abduzeedo

Artifact from the MuFo: Branding and Visual Identity for Photography Museum article on AbduzeedoArtifact from the MuFo: Branding and Visual Identity for Photography Museum article on Abduzeedo

Artifact from the MuFo: Branding and Visual Identity for Photography Museum article on Abduzeedo

Artifact from the MuFo: Branding and Visual Identity for Photography Museum article on Abduzeedo

Artifact from the MuFo: Branding and Visual Identity for Photography Museum article on Abduzeedo

For more information make sure to check out podpunkt.pl /​​​​​​​ superskrypt.pl ​​​​​​​

The Feature Trap: Why Feature Centricity Is Harming Your Product

Original Source: https://smashingmagazine.com/2024/01/feature-centricity-harming-product/

Most product teams think in terms of features. Features are easy to brainstorm and write requirement docs for, and they fit nicely into our backlogs and ticketing systems. In short, thinking in terms of features makes it easy to manage the complex task of product delivery.

However, we know that the best products are more than the sum of their parts, and sometimes, the space between the features is as important as the features themselves. So, what can we do to improve the process?

The vast majority of product teams are organized around delivering features — new pieces of functionality that extend the capabilities of the product. These features will often arise from conversations the company is having with prospective buyers:

“What features are important to you?”
“What features are missing from your current solution?”
“What features would we need to add in order to make you consider switching from your existing provider to us?” and so on.

The company will then compile a list of the most popular feature requests and will ask the product team to deliver them.

For most companies, this is what customer centricity looks like; asking customers to tell them what they want — and then building those features into the product in the hope they’ll buy — becomes of key importance. This is based on the fundamental belief that people buy products primarily for the features so we assemble our roadmaps accordingly.

We see this sort of thinking with physical products all the time. For instance, take a look at the following Amazon listing for one of the top-rated TV sets from last year. It’s like they hurled up the entire product roadmap directly onto the listing!

Now, of course, if you’re a hardcore gamer with very specific requirements, you might absolutely be looking for a TV with “VRR, ALLM, and eARC as specified in HDMI2.1, plus G-Sync, FreeSync, Game Optimizer, and HGiG.” But for me? I don’t have a clue what any of those things mean, and I don’t really care. Instead, I’ll go to a review site where they explain what the product actually feels like to use in everyday life. The reviewers will explain how good the unboxing experience is. How sturdy the build is. How easy it is to set up. They’ll explain that the OS is really well put together and easy to navigate, the picture quality is probably the best on the market, and the sound, while benefiting from the addition of a quality sound bar, is very clear and understandable. In short, they’ll be describing the user experience.

The ironic thing is that when I talk to most founders, product managers, and engineers about how they choose a TV, they’ll say exactly the same thing. And yet, for some reason, we struggle to take that personal experience and apply it to our own users!

Tip: As a fun little trick, next time you find yourself arguing about features over experience, ask people to get out their phones. I bet that the vast majority of folks in the room will have an iPhone, despite Samsung and Google phones generally having better cameras, more storage, better screens, and so on. The reason why iPhones have risen in dominance (if we ignore the obvious platform lock-in) is because, despite perhaps not having the best feature set on the market, they feel so nice to use.

Seeing Things From The Users’ Perspective

While feature-centric thinking is completely understandable, it misses a whole class of problems. The features in and of themselves might look good on paper and work great in practice, but do they mesh together to form a convincing whole? Or is the full experience a bit of a mess?

All the annoying bumps, barriers, and inconsistencies that start accruing around each new feature, if left unsolved, can limit the amount of value users can extract from the product. And if you don’t effectively identify and remove these barriers in a deliberate and structured way, any additional functionality will simply add to the problem.

If users are already struggling to extract value from existing features, how do you expect them to extract any additional value you might be adding to the product?

“As a product manager, it’s natural to want to offer as many features as possible to your customers. After all, you want to provide value, right? But what happens when you offer too many features? Your product becomes bloated, convoluted, and difficult to use.”
— “Are Too Many Features Hurting Your Product?”

These barriers and inconsistencies are usually the result of people not thinking through the user experience. And I don’t mean user experience in some abstract way. I mean literally walking through the product step-by-step as though you’d never seen it before — sometimes described as having a “beginner’s mind” mdash; and considering the following questions:

Is it clear what value this product delivers and how I can get that value?
If I were a new user, would the way the product is named and structured make sense to me?
Can I easily build up a mental model of where everything is and how the product works?
Do I know what to do next?
How is this going to fit into my existing workflow?
What’s getting in my way and slowing me down?

While approaching things with a beginner’s mind sounds easy, it’s actually a surprisingly hard mindset for people to adopt — letting go of everything they know (or think they know) about their product, market, and users. Instead, their position as a superuser tends to cloud their judgment: believing that because something is obvious to them (something that they have created and have been working on for the past two years), it will be obvious to a new user who has spent less than five minutes with the product. This is where usability testing (a UX research method that evaluates whether users are able to use a digital product efficiently and effectively) should normally “enter the stage.”

The issue with trying to approach things with a beginner’s mind is also often exacerbated by “motivated reasoning,” the idea that we view things through the lens of what we want to be true, rather than what is true. To this end, you’re much more likely to discount feedback from other people if that feedback is going to result in some negative outcome, like having to spend extra time and money redesigning a user flow when you’d rather be shipping that cool new feature you came up with last week.

I see this play out in usability testing sessions all the time. The first subject comes in and struggles to grasp a core concept, and the team rolls their eyes at the incompetence of the user. The next person comes in and has the same experience, causing the team to ask where you found all these stupid users. However, as the third, fourth, and fifth person comes through and experiences the same challenge, “lightbulbs” slowly start forming over the team members’ heads:

“Maybe this isn’t the users’ fault after all? Maybe we’ve assumed a level of knowledge or motivation that isn’t there; maybe it’s the language we’ve used to describe the feature, or maybe there’s something in the way the interface has been designed that is causing this confusion?”

These kinds of insights can cause teams to fundamentally pivot their thinking. But this can also create a huge amount of discomfort and cognitive dissonance — realizing that your view of the world might not be entirely accurate. As such, there’s a strong motivation for people to avoid these sorts of realizations, which is why we often put so little effort (unfortunately) into understanding how our users perceive and use the things we create.

Developing a beginner’s mind takes time and practice. It’s something that most people can cultivate, and it’s actually something I find designers are especially good at — stepping into other people’s shoes, unclouded by their own beliefs and biases. This is what designers mean when they talk about using empathy.

Towards A Two-Tier Process (Conclusion)

We obviously still need to have “feature teams.” Folks who can understand and deliver the new capabilities our users request (and our business partners demand). While I’d like to see more thought and validation when it comes to feature selection and creation, it’s often quicker to add new features to see if they get used than to try and use research to give a definitive answer.

As an example, I’m working with one founder at the moment who has been going around houses with their product team for months about whether a feature would work. He eventually convinced them to give it a try — it took four days to push out the change, and they got the feedback they needed almost instantly.

However, as well as having teams focused on delivering new user value, we also need teams who are focused on helping unlock and maximize existing user value. These teams need to concentrate on outcomes over outputs; so, less deliver X capability in Y sprints than deliver X improvement by Y date. To do this, these teams need to have a high level of agency. This means taking them out of the typical feature factory mindset.

The teams focusing on helping unlock and maximize existing user value need to be a little more cross-disciplinary than your traditional feature team. They’re essentially developing interventions rather than new capabilities — coming up with a hypothesis and running experiments rather than adding bells and whistles. “How can we improve the onboarding experience to increase activation and reduce churn?” Or, “How can we improve messaging throughout the product so people have a better understanding of how it works and increase our North Star metric as a result?”

There’s nothing radical about focusing on outcomes over outputs. In fact, this way of thinking is at the heart of both the Lean Startup movement and the Product Led Growth. The problem is that while this is seen as received wisdom, very few companies actually put it into practice (although if you ask them, most founders believe that this is exactly what they do).

Put simply, you can’t expect teams to work independently to deliver “outcomes” if you fill their their calendar with output work.

So this two-tier system is really a hack, allowing you to keep sales, marketing, and your CEO (and your CEO’s partner) happy by delivering a constant stream of new features while spinning up a separate team who can remove themselves from the drum-beat of feature delivery and focus on the outcomes instead.

Further Reading

“Why Too Many Features Can Ruin a Digital Product Before It Begins” (Komodo Digital)
Digital products are living, ever-evolving things. So, why do so many companies force feature after feature into projects without any real justification? Let’s talk about feature addiction and how to avoid it.
“Are Too Many Features Hurting Your Product?” (FAQPrime)
As a product manager, it’s natural to want to offer as many features as possible to your customers. After all, you want to provide value, right? But what happens when you offer too many features? Your product becomes bloated, convoluted, and difficult to use. Let’s take a closer look at what feature bloat is, why it’s a problem, and how you can avoid it.
“Twelve Signs You’re Working in a Feature Factory,” John Cutler
The author started using the term Feature Factory when a software developer friend complained that he was “just sitting in the factory, cranking out features, and sending them down the line.” This article was written in 2016 and still holds its ground today. In 2019 there appeared a newer version of it (“Twelve signs You’re Working in a Feature Factory — Three Years Later”).
“What Is The Agile Methodology?,” (Atlassian)
The Agile methodology is a project management approach that involves breaking the project into phases and emphasizes continuous collaboration and improvement. Teams follow a cycle of planning, executing, and evaluating.
“Problem Statement vs Hypothesis — Which ­­Is More Important?,” Sadie Neve
When it comes to experimentation and conversion rate optimization (CRO), we often see people relying too much on their instincts. But in reality, nothing in experimentation is certain until tested. This means experimentation should be approached like a scientific experiment that follows three core steps: identify a problem, form a hypothesis, and test that hypothesis.
“The Build Trap,” Melissa Perri (Produx Labs)
The “move fast and break things” mantra seems to have taken the startup world by storm since Facebook made it their motto a few years ago. But there is a serious flaw with this phrase, and it’s that most companies see this as an excuse to stop analyzing what they intend to build and why they should build it — those companies get stuck in what I call “The Build Trap.”
“What Is Product-led Growth?” (PLG Collective)
We are in the middle of a massive shift in the way people use and buy software. It’s been well over a decade since Salesforce brought software to the cloud. Apple put digital experiences in people’s pockets back in 2009 with the first iPhone. And in the years since the market has been flooded with consumer and B2B products that promise to meet just about every need under the sun.
The Lean Startup
The Lean Startup isn’t just about how to create a more successful entrepreneurial business. It’s about what we can learn from those businesses to improve virtually everything we do.
“Usability Testing — The Complete Guide,” Daria Krasovskaya and Marek Strba
Usability testing is the ultimate method of uncovering any type of issue related to a system’s ease of use, and it truly is a must for any modern website or app owner.
“The Value of Great UX,” Jared Spool
How can we show that a great user experience produces immense value for the organization? We can think of experience as a spectrum, from extreme frustration to delight. In his article, Jared will walk you through how our work as designers is able to transform our users’ experiences from being frustrated to being delighted.
“Improving The Double Diamond Design Process,” Andy Budd (Smashing Magazine)
The so-called “Double Diamond” is a great way of visualizing an ideal design process, but it’s just not the way most companies deliver new projects or services. The article proposes a new “Double Diamond” idea that better aligns with the way work actually gets done and highlights the place where design has the most leverage.
“Are We Moving Towards a Post-Agile Age?,” Andy Budd
Agile has been the dominant development methodology in our industry for a while now. While some teams are just getting to grips with Agile, others have extended it to the point that it’s no longer recognizable as Agile; in fact, many of the most progressive design and development teams are Agile only in name. What they are actually practicing is something new, different, and innately more interesting — something I’ve been calling Post-Agile thinking.

A Simple Guide To Retrieval Augmented Generation Language Models

Original Source: https://smashingmagazine.com/2024/01/guide-retrieval-augmented-generation-language-models/

Suppose you ask some AI-based chat app a reasonably simple, straightforward question. Let’s say that app is ChatGPT, and the question you ask is right in its wheelhouse, like, “What is Langchain?” That’s really a softball question, isn’t it? ChatGPT is powered by the same sort of underlying technology, so it ought to ace this answer.

So, you type and eagerly watch the app spit out conversational strings of characters in real-time. But the answer is less than satisfying.

In fact, ask ChatGPT — or any other app powered by language models — any question about anything recent, and you’re bound to get some sort of response along the lines of, “As of my last knowledge update…” It’s like ChatGPT fell asleep Rumplestiltskin-style back in January 2022 and still hasn’t woken up. You know how people say, “You’d have to be living under a rock not to know that”? Well, ChatGPT took up residence beneath a giant chunk of granite two years ago.

While many language models are trained on massive datasets, data is still data, and data becomes stale. You might think of it like Googling “CSS animation,” and the top result is a Smashing Magazine article from 2011. It might still be relevant, but it also might not. The only difference is that we can skim right past those instances in search results while ChatGPT gives us some meandering, unconfident answers we’re stuck with.

There’s also the fact that language models are only as “smart” as the data used to train them. There are many techniques to improve language model’s performance, but what if language models could access real-world facts and data outside their training sets without extensive retraining? In other words, what if we could supplement the model’s existing training with accurate, timely data?

This is exactly what Retrieval Augmented Generation (RAG) does, and the concept is straightforward: let language models fetch relevant knowledge. This could include recent news, research, new statistics, or any new data, really. With RAG, a large language model (LLM) is able to retrieve “fresh” information for more high-quality responses and fewer hallucinations.

But what exactly does RAG make available, and where does it fit in a language chain? We’re going to learn about that and more in this article.

Understanding Semantic Search

Unlike keyword search, which relies on exact word-for-word matching, semantic search interprets a query’s “true meaning” and intent — it goes beyond merely matching keywords to produce more results that bear a relationship to the original query.

For example, a semantic search querying “best budget laptops” would understand that the user is looking for “affordable” laptops without querying for that exact term. The search recognizes the contextual relationships between words.

This works because of text embeddings or mathematical representations of meaning that capture nuances. It’s an interesting process of feeding a query through an embedded model that, in turn, converts the query into a set of numeric vectors that can be used for matching and making associations.

The vectors represent meanings, and there are benefits that come with it, allowing semantic search to perform a number of useful functions, like scrubbing irrelevant words from a query, indexing information for efficiency, and ranking results based on a variety of factors such as relevance.

Special databases optimized for speed and scale are a strict necessity when working with language models because you could be searching through billions of documents. With a semantic search implementation that includes test embedding, storing and querying high-dimensional embedding data is much more efficient, producing quick and efficient evaluations on queries against document vectors across large datasets.

That’s the context we need to start discussing and digging into RAG.

Retrieval Augmented Generation

Retrieval Augmented Generation (RAG) is based on research produced by the Meta team to advance the natural language processing capabilities of large language models. Meta’s research proposed combining retriever and generator components to make language models more intelligent and accurate for generating text in a human voice and tone, which is also commonly referred to as natural language processing (NLP).

At its core, RAG seamlessly integrates retrieval-based models that fetch external information and generative model skills in producing natural language. RAG models outperform standard language models on knowledge-intensive tasks like answering questions by augmenting them with retrieved information; this also enables more well-informed responses.

You may notice in the figure above that there are two core RAG components: a retriever and a generator. Let’s zoom in and look at how each one contributes to a RAG architecture.

Retriever

We already covered it briefly, but a retriever module is responsible for finding the most relevant information from a dataset in response to queries and makes that possible with the vectors produced by text embedding. In short, it receives the query and retrieves what it evaluates to be the most accurate information based on a store of semantic search vectors.

Retrievers are models in and of themselves. But unlike language models, retrievers are not in the business of “training” or machine learning. They are more of an enhancement or an add-on that provides additional context for understanding and features for fetching that information efficiently.

That means there are available options out there for different retrievers. You may not be surprised that OpenAI offers one, given their ubiquity. There’s another one provided by Cohere as well as a slew of smaller options you can find in the Hugging Face community.

Generator

After the retriever finds relevant information, it needs to be passed back to the application and displayed to the user. Or what’s needed is a generator capable of converting the retrieved data into human-readable content.

What’s happening behind the scenes is the generator accepts the embeddings it receives from the retriever, mashes them together with the original query, and passes through the trained language model for an NLP pass on the way to becoming generated text.

The entire tail end of that process involving the language model and NLP is a process in its own right and is something I have explained in greater detail in another Smashing Magazine article if you are curious about what happens between the generator and final text output.

RAG Full View

Pulling everything together, a complete RAG flow goes like this:

A query is made.
The query is passed to the RAG model.
The RAG model encodes the query into text embeddings that are compared to a dataset of information.
The RAG’s retriever decides the most relevant information with its semantic search abilities and converts it into vector embeddings.
The RAG’s retriever sends the parsed embeddings to the generator.
The generator accepts the embeddings and combines them with the original query.
The generator passes its work off to the language model to produce natural-sounding content presented to the user.

LLM Hallucinations And Knowledge Limitations

We opened this article up by describing “hallucinations” in LLMs’ incorrect responses or something along the lines of “I don’t know, but here’s what I do know.” The LLM will “make stuff up” because it simply doesn’t have updated information to respond with.

Let’s revisit the first query we used to kick off this article — “What is LangChain?” — and compare responses from the Vicuna and GPT-4 language models:

Here’s the transcription for the second query using OpenAI’s GPT-4 for posterity:

“LangChain is a blockchain-based decentralized translation platform. It’s designed to connect translators and those seeking translation services directly without the need for intermediaries. This system uses artificial intelligence and blockchain technologies to provide high-quality, efficient translation services. LangChain also has a token-based economy, where users can earn and spend tokens within the LangChain ecosystem.”

Both Vicuna and GPT-4 refer to LangChain as a blockchain platform. Blockchain is a technology that stores data in a decentralized manner using chained blocks, so the models’ responses sound plausible given the “chain” in the name. However, LangChain is not actually a blockchain-based technology.

This is a prime example demonstrating how LLMs will fabricate responses that may seem believable at first glance but are incorrect. LLMs are designed to predict the next “plausible” tokens in a sequence, whether those are words, subwords, or characters. They don’t inherently understand the full meaning of the text. Even the most advanced models struggle to avoid made-up responses, especially for niche topics they lack knowledge about.

Let’s look at another example by querying: “What is the most preferred framework used by developers for building applications leveraging large language models?”

While Vicuna offers a couple of reasonable starting points for answering the question, the frameworks it refers to have limitations for efficiency and scalability in production-level applications that use LLMs. That could quite possibly send a developer down a bad path. And as bad as that is, look at the GPT-4 response that changes topics completely by focusing on LLVM, which has nothing to do with LLMs.

What if we refine the question, but this time querying different language models? This time, we’re asking: “What is the go-to framework developed for developers to seamlessly integrate large language models into their applications, focusing on ease of use and enhanced capabilities?”

Honestly, I was expecting the responses to refer to some current framework, like LangChain. However, the GPT-4 Turbo model suggests the “Hugging Face” transformer library, which I believe is a great place to experiment with AI development but is not a framework. If anything, it’s a place where you could conceivably find tiny frameworks to play with.

Meanwhile, the GPT-3.5 Turbo model produces a much more confusing response, talking about OpenAI Codex as a framework, then as a language model. Which one is it?

We could continue producing examples of LLM hallucinations and inaccurate responses and have fun with the results all day. We could also spend a lot of time identifying and diagnosing what causes hallucinations. But we’re here to talk about RAG and how to use it to prevent hallucinations from happening in the first place. The Master of Code Global blog has an excellent primer on the causes and types of LLM hallucinations with lots of useful context if you are interested in diving deeper into the diagnoses.

Integrating RAG With Language Models

OK, so we know that LLMs sometimes “hallucinate” answers. We know that hallucinations are often the result of outdated information. We also know that there is this thing called Retrieval Augmented Generation that supplements LLMs with updated information.

But how do we connect RAG and LLMs together?

Now that you have a good understanding of RAG and its benefits, we can dive into how to implement it yourself. This section will provide hands-on examples to show you how to code RAG systems and feed new data into your LLM.

But before jumping right into the code, you’ll need to get a few key things set up:

Hugging Face
We’ll use this library in two ways. First, to choose an embedding model from the model hub that we can use to encode our texts, and second, to get an access token so we can download the Llama-2 model. Sign up for a free Hugging Face in preparation for the work we’ll cover in this article.
Llama-2
Meta’s powerful LLM will be our generator model. Request access via Meta’s website so we can integrate Llama-2 into our RAG implementation.
LlamaIndex
We’ll use this framework to load our data and feed it into Llama-2.
Chroma
We’ll use this embedding database for fast vector similarity search and retrieval. This is actually where we can store our index.

With the key tools in place, we can walk through examples for each phase: ingesting data, encoding text, indexing vectors, and so on.

Install The Libraries

We need to install the RAG libraries we identified, which we can do by running the following commands in a new project folder:

# Install essential libraries for our project
!pip install llama-index transformers accelerate bitsandbytes –quiet
!pip install chromadb sentence-transformers pydantic==1.10.11 –quiet

Next, we need to import specific modules from those libraries. There are quite a few that we want, like ChromaVectorStore and HuggingFaceEmbedding for vector indexing and embeddings capabilities, storageContext and chromadb to provide database and storage functionalities, and even more for computations, displaying outputs, loading language models, and so on. This can go in a file named app.py at the root level of your project.

## app.py

## Import necessary libraries
from llama_index import VectorStoreIndex, download_loader, ServiceContext
from llama_index.vector_stores import ChromaVectorStore
from llama_index.storage.storage_context import StorageContext
from llama_index.embeddings import HuggingFaceEmbedding
from llama_index.response.notebook_utils import display_response
import torch
from transformers import BitsAndBytesConfig
from llama_index.prompts import PromptTemplate
from llama_index.llms import HuggingFaceLLM
from IPython.display import Markdown, display
import chromadb
from pathlib import Path
import logging
import sys

Provide Additional Context To The Model

The data we will leverage for our language model is a research paper titled “Enhancing LLM Intelligence with ARM-RAG: Auxiliary Rationale Memory for Retrieval Augmented Generation” (PDF) that covers an advanced retrieval augmentation generation approach to improve problem-solving performance.

We will use the download_loader() module we imported earlier from llama_index to download the PDF file:

PDFReader = download_loader(“PDFReader”)
loader = PDFReader()
documents = loader.load_data(file=Path(‘/content/ARM-RAG.pdf’))

Even though this demonstration uses a PDF file as a data source for the model, that is just one way to supply the model with data. For example, there is Arxiv Papers Loader as well as other loaders available in the LlamaIndex Hub. But for this tutorial, we’ll stick with loading from a PDF. That said, I encourage you to try other ingestion methods for practice!

Now, we need to download Llama-2, our open-source text generation model from Meta. If you haven’t already, please set up an account with Meta and have your access token available with read permissions, as this will allow us to download Llama-2 from Hugging Face.

# huggingface api token for downloading llama2
hf_token = “YOUR Access Token”

To fit Llama-2 into constrained memory, like in Google Colab, we’ll configure 4-bit quantization to load the model at a lower precision.

quantization_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_compute_dtype=torch.float16,
bnb_4bit_quant_type=”nf4″,
bnb_4bit_use_double_quant=True,
)

Google Colab is where I typically do most of my language model experiments. We’re shrinking the language model down with that last snippet so it’s not too large for Colab to support.

Next, we need to initialize HuggingFaceLLM to set up our model for generating text on the fly:

llm = HuggingFaceLLM(
model_name=”meta-llama/Llama-2-7b-chat-hf”,
tokenizer_name=”meta-llama/Llama-2-7b-chat-hf”,
query_wrapper_prompt=PromptTemplate(“<s> [INST] {query_str} [/INST] “),
context_window=3900,
model_kwargs={“token”: hf_token, “quantization_config”: quantization_config},
tokenizer_kwargs={“token”: hf_token},
device_map=”auto”,
)

Note that what’s happening here is that we’ve defined our model (Llama-2), but a specific subset of it is designed for a chat UI (Llama-2-Chat). Even though this is what we are using for this article, you can also pick from other text generation models in the Hugging Face hub, such as Neural-chat from Intel.

Compare Outputs

OK, so we have the model loaded and configured. Let’s try it out by asking it questions that it could reasonably answer using the content we provided via the ARM-RAG PDF file to see how the model responds without any context. The following code is for displaying the text output without the PDF providing data to the model for comparison sake:

from IPython.display import HTML, display

# Assuming resp contains the response
resp = llm.complete(“What is ARM-RAG?”)

# Using HTML with inline CSS for styling (gray color, smaller font size)
html_text = f'<p style=”color: #1f77b4; font-size: 14px;”><b>{resp}</b></p>’

Let’s see the response:

display(HTML(html_text))

ARM-RAG (ARM Register-Allocating Generator) is a register allocation algorithm used in compiler design. nobody knows who invented it, but it was first described in a 1995 paper by Mark H. T. Kim and Andrew Y. Chen. The ARM-RAG algorithm is a register-allocating generator, meaning it is a heuristic algorithm that takes a program as input and generates a register allocation plan for it. The goal of the algorithm is to assign registers to variables in a way that minimizes the number of spills (i.e., loads or stores to/from memory) while satisfying all the memory access constraints of the program. The ARM-RAG algorithm works by first partitioning the program into a set of basic blocks, and then iteratively allocating registers to variables within each basic block. The algorithm uses a cost function to evaluate the quality of each allocation, and it selects the allocation with the lowest cost. The cost function takes into account factors such as the distance between the variable and the current register, the distance between the variable and the next allocation, and the number of spills that would be required to allocate the variable. ARM-RAG

Obviously, without any context from the PDF file, the model generates a hallucinated response that describes some unrelated algorithm. To fix this, we need to provide relevant passages from the paper to ground Llama-2’s responses. We’ll encode the document into embeddings and index them for retrieval; then, when we query, we can feed LLama-2 relevant retrieved passages to steer it toward accurate, on-topic responses based on the contents of the PDF file.

First, we need to create a client to interact with our ChromaDB database and a new collection that will hold our vector index.

# create client and a new collection
chroma_client = chromadb.EphemeralClient()
chroma_collection = chroma_client.create_collection(“firstcollection”)

Then we need to set up the HuggingFaceEmbedding class with the specified model name for embedding the text into vectors:

# Load the embedding model
embed_model = HuggingFaceEmbedding(model_name=”BAAI/bge-base-en-v1.5″)

This initializes HuggingFaceEmbedding, passing the name of the pre-trained model we want to use, BAAI/bge-base-en-v1.5. There are other options, of course.

Now, we can set up the vector store and use it to index the embedded document vectors:

# set up ChromaVectorStore and load in data
vector_store = ChromaVectorStore(chroma_collection=chroma_collection)
storage_context = StorageContext.from_defaults(vector_store=vector_store)
service_context = ServiceContext.from_defaults(llm=llm, embed_model=embed_model)
index = VectorStoreIndex.from_documents(
documents, storage_context=storage_context, service_context=service_context
)

This creates a ChromaVectorStore connected to our collection, defines the storage and service contexts, and generates a VectorStoreIndex from the loaded documents using the embedding model. The index is what allows us to quickly find relevant passages for a given query to augment the quality of the model’s response.

We should also establish a way for the model to summarize the data rather than spitting everything out at once. A SummaryIndex offers efficient summarization and retrieval of information:

summary_index = SummaryIndex.from_documents(documents, service_context=service_context)

Earlier, the model hallucinated when we queried it without the added context from the PDF file. Now, let’s ask the same question, this time querying our indexed data:

#Define your query
query=”what is ARM-RAG?”

from llama_index.embeddings.base import similarity
query_engine =index.as_query_engine(response_mode=”compact”)
response = query_engine.query(query)
from IPython.display import HTML, display

# Using HTML with inline CSS for styling (blue color)
html_text = f'<p style=”color: #1f77b4; font-size: 14px;”><b>{response}</b></p>’
display(HTML(html_text))

Here’s the output:

Final Response: Based on the context information provided, ARM-RAG is a system that utilizes Neural Information Retrieval to archive reasoning chains derived from solving grade-school math problems. It is an Auxiliary Rationale Memory for Retrieval Augmented Generation, which aims to enhance the problem-solving capabilities of Large Language Models (LLMs). The system surpasses the performance of a baseline system that relies solely on LLMs, demonstrating the potential of ARM-RAG to improve problem-solving capabilities.

Correct! This response is way better than the one we saw earlier — no hallucinations here.

Since we’re using the chat subset of the Llama-2 model, we could have a back-and-forth conversation with the model about the content of the PDF file with follow-up questions. That’s because the indexed data supports NLP.

chat_engine = index.as_chat_engine(chat_mode=”condense_question”, verbose=True)
response = chat_engine.chat(“give me real world examples of apps/system i can build leveraging ARM-RAG?”)
print(response)

This is the resulting output:

Querying with: What are some real-world examples of apps or systems that can be built leveraging the ARM-RAG framework, which was discussed in our previous conversation?
Based on the context information provided, the ARM-RAG framework can be applied to various real-world examples, including but not limited to:

1. Education: ARM-RAG can be used to develop educational apps that can help students learn and understand complex concepts by generating explanations and examples that can aid in their understanding.

2. Tutoring: ARM-RAG can be applied to tutoring systems that can provide personalized explanations and examples to students, helping them grasp difficult concepts more quickly and effectively.

3. Customer Service: ARM-RAG can be utilized in chatbots or virtual assistants to provide customers with detailed explanations and examples of products or services, enabling them to make informed decisions.

4. Research: ARM-RAG can be used in research environments to generate explanations and examples of complex scientific concepts, enabling researchers to communicate their findings more effectively to a broader audience.

5. Content Creation: ARM-RAG can be applied to content creation systems that can generate explanations and examples of complex topics, such as news articles, blog posts, or social media content, making them more engaging and easier

Try asking more questions! Now that the model has additional context to augment its existing dataset, we can have a more productive — and natural — interaction.

Additional RAG Tooling Options

The whole point of this article is to explain the concept of RAG and demonstrate how it can be used to enhance a language model with accurate and updated data.

Chroma and LlamaIndex were the main components of the demonstrated RAG approach, but there are other tools for integrating RAG with language models. I’ve prepared a table that outlines some popular options you might consider trying with your own experiments and projects.

RAG
Type of System
Capabilities
Integrations
Documentation / Repo

Weaviate
Vector Database
Vector & Generative search
LlamaIndex, LangChain, Hugging Face, Cohere, OpenAI, etc.
DocumentationGitHub

Pinecone
Vector Database
Vector search, NER-Powered search, Long-term memory
OpenAI, LangChain, Cohere, Databricks
DocumentationGitHub

txtai
Embeddings Database
Semantic graph & search, Conversational search
Hugging face models
DocumentationGitHub

Qdrant
Vector Database
Similarity image search, Semantic search, Recommendations
LangChain, LlamaIndex, DocArray, Haystack, txtai, FiftyOne, Cohere, Jina Embeddings, OpenAI
DocumentationGitHub

Haystack
Framework
QA, Table QA, Document search, Evaluation
Elasticsearch, Pinecone, Qdrant, Weaviate, vLLM, Cohere
DocumentationGitHub

Ragchain
Framework
Reranking, OCR loaders
Hugging Face, OpenAI, Chroma, Pinecone
DocumentationGitHub

metal
Vector Database
Clustering, Semantic search, QA
LangChain, LlamaIndex
DocumentationGitHub

Conclusion

In this article, we examined examples of language models producing “hallucinated” responses to queries as well as possible causes of those hallucinations. At the end of the day, a language model’s responses are only as good as the data it provided, and as we’ve seen, even the most widely used models consist of outdated information. And rather than admit defeat, the language model spits out confident guesses that could be misconstrued as accurate information.

Retrieval Augmented Generation is one possible cure for hallucinations.

By embedding text vectors pulled from additional sources of data, a language model’s existing dataset is augmented with not only new information but the ability to query it more effectively with a semantic search that helps the model more broadly interpret the meaning of a query.

We did this by registering a PDF file with the model that contains content the model could use when it receives a query on a particular subject, in this case, “Enhancing LLM Intelligence with ARM-RAG: Auxiliary Rationale Memory for Retrieval Augmented Generation.”

This, of course, was a rather simple and contrived example. I wanted to focus on the concept of RAG more than its capabilities and stuck with a single source of new context around a single, specific subject so that we could easily compare the model’s responses before and after implementing RAG.

That said, there are some good next steps you could take to level up your understanding:

Consider using high-quality data and embedding models for better RAG performance.
Evaluate the model you use by checking Vectara’s hallucination leaderboard and consider using their model instead. The quality of the model is essential, and referencing the leaderboard can help you avoid models known to hallucinate more often than others.
Try refining your retriever and generator to improve results.

My previous articles on LLM concepts and summarizing chat conversations are also available to help provide even more context about the components we worked with in this article and how they are used to produce high-quality responses.

References

LlamaIndex documentation
ChromaDB documentation
Metas Llama-2 access
ARM-RAG research paper

CSS Blurry Shimmer Effect

Original Source: https://smashingmagazine.com/2024/01/css-blurry-shimmer-effect/

Imagine box-shadow but for a blur effect, where the backdrop of an element is blurred around that element, gradually decreasing the blur’s strength. I came up with the idea while trying to improve the contrast of a popup over a dark area where a box-shadow for the popup won’t make much sense, design-wise. I then thought, well, what other ways might create a good contrast effect? And so suddenly, the idea of a gradual blur effect around the object came to me.

See the Pen Faded Outer Box Backdrop Blur [forked] by Yair Even Or.

It would be awesome if we had a box-blur property or perhaps some sort of blur keyword we could set on box-shadow the way we do for inset shadows. Unfortunately, CSS has no such property. But because CSS is awesome and flexible, we can still get the effect by combining a few CSS features and hack it through.

What I’m going to show you from here on out is the thought process I took to create the effect. Sometimes, I find it easier to know what’s coming up rather than meandering through a narrative of twists and turns. So, for those of you who are like me and want to jump straight into the process, this was my approach.

Start With The Markup

The effect is approached in a way that it is applied to the ::before pseudo-element of some element, say some popup/dialog/popover/tooltip. Those are the common “targets” for this sort of effect. I think using a pseudo-element is a good approach here because it means we could technically scope the styles to the pseudo-element and re-purpose the effect on other elements without any HTML changes.

<!– This is literally it for this demo –>
<div></div>

You can give the element a class, whatever dimensions you like, insert content and other child elements within it, or use a completely different element. The HTML isn’t the main ingredient for the secret sauce we’re making.

Position The Pseudo-Element

We want the ::before pseudo-element to occupy the entire area of the <div> element we’re using for this specific demo. Not only do we want it to cover the entire area, but even overflow it because that establishes the visible area, which holds the blur effect, so it will extend outwards.

::before {
content: ”;

/* Make sure the parent element is at least relatively positioned to contain the pseudo-element. */
position: absolute;

/* The blur size should be anything below 0 so it will extend to the outside. */
inset: -100px;

/* This layer is positioned between the parent element and page background. */
/* Make sure this value is one below the z-index of the parent element. */
z-index: -1;
}

The code comments spell out the key pieces. An empty string has to be set for the content property so the ::before will be rendered, then we take it out of the document flow by giving it absolute positioning. This allows us to inset the element’s position and is ultimately setting the blur effect directions as we would on the box-shadow property — only we’re using inset to control its size. We want a negative inset value, where the effect extends further the lower the value gets.

Until now, we’ve set the foundation for the effect. There’s nothing really to see just yet. Now, the fun begins!

Masking With Transparent Gradients

Gradients are technically images — generated by the browser — which can be used as CSS masks to hide parts of an element to create various shapes. You may have seen a few related Smashing Magazine articles where CSS masking has been showcased, such as this one by Temani Afif.

Transparency is the key thing when it comes to masking with gradients. Transparency allows us to gradually hide portions of an element in a way that creates the illusion of fading in or out.

That’s perfect in this case because we want the effect to be stronger, closer to the object, and fade in intensity as it gets further away.

We’ll use two gradients: one that goes horizontally and another that goes vertically. I chose this route because it mimics a rough rectangle shape that fades out towards the edges.

As I said, transparency is key. Both gradients start transparent, then transition to black until just before the end, where they go back to transparent to fade things out. Remember, these gradients are masks rather than background images, so they are declared on the mask property, which controls which pixels should be rendered and their opacity.

mask:
linear-gradient(to top, transparent 0%, black 25% 75%, transparent 100%),
linear-gradient(to left, transparent 0%, black 25% 75%, transparent 100%);

See the Pen Basic Gradient Mask [forked] by Yair Even Or.

The vertical gradient (to top) creates a fade from transparent at the bottom to black in the middle, then back to transparent at the top.
The horizontal gradient (to left) produces a fade from transparent on the right to black in the middle, then back to transparent on the left.

This dual-gradient approach positions the black regions, so they merge, creating the rough baseline of a rectangular shape that will be refined in the next step. The mask property is best declared as first prefixed and then un-prefixed to cover more browsers’ support:

-webkit-mask:
linear-gradient(to top, transparent 0%, black 25% 75%, transparent 100%),
linear-gradient(to left, transparent 0%, black 25% 75%, transparent 100%);
mask:
linear-gradient(to top, transparent 0%, black 25% 75%, transparent 100%),
linear-gradient(to left, transparent 0%, black 25% 75%, transparent 100%);

Refining Using The mask-composite Property

The mask-composite property is part of the CSS Masking Module and enables pixel-wise control over the blending of masked content, allowing for intricate compositions.

The source-in value of this property is very useful for the effect we are after because it tells the browser to only retain the overlapping areas of the mask, so only pixels that contain both (mentioned above) gradients will get rendered. This locks in a rectangle shape, which can then be applied on any DOM element that has none-to-moderately curved corners (border-radius).

Gradually Blurring The Backdrop

Now that we have a mask to work with, all we need to do is use it. The backdrop-filter CSS property can blur anything that is rendered “behind” an element using the blur() function:

::before {
/* etc. */

backdrop-filter: blur(10px);
}

The larger the value, the more intense the blur. I’m using 10px arbitrarily. In fact, we can variablize this stuff later to make the implementation even more flexible and easily configurable.

But wait! As it turns out, Safari requires a vendor-prefixed version of backdrop-filter to get it working there:

::before {
/* etc. */

-webkit-backdrop-filter: blur(10px); /* Required for Safari */
backdrop-filter: blur(10px);
}

Note: It’s preferred to declare prefixed properties before the unprefixed variant so they serve as a fallback for browsers that don’t (yet) support them or their implementation is different.

A Touch of Synergistic Shadow

I think adding a slight semi-opaque black box-shadow that covers the blur area gives the effect a little extra depth. The only thing is that you’ll want to add it to the element itself rather than it’s ::before pseudo:

div {
box-shadow: 0 0 40px #00000099;
}

That’s totally optional, though.

Bringing Everything Together

Here’s how the CSS comes out when we combine everything together.

/* This can be set on the ::before pseudo of the element it is applied to. */
::before {
content: ”;

/* This layer is positioned between some element and its background. */
position: absolute;

/* This should not affect the contents of the container. */
z-index: -1;

/* The blur size should be anything below 0 so it will extend to the outside. */
inset: -100px;

/* The blur effect */
-webkit-backdrop-filter: blur(10px); /* Required for safari */
backdrop-filter: blur(10px);

/* A mask fades the blur effect, so it gets weaker. */
/* towards the edges, further from the container box. */
/* (The fill color is irrelevant, so “red” is used as it’s the shortest color name.) */
mask:
linear-gradient(
to top,
transparent 0%,
red 100px calc(100% – 100px),
transparent 100%),
linear-gradient(
to left,
transparent 0%,
red 100px calc(100% – 100px),
transparent 100%);

/* This merges the masks above so only the overlapping pixels are rendered. */
/* This creates the illusion of a fade-out mask. */
mask-composite: intersect;
-webkit-mask-composite: source-in; /* Required for Safari */
}

The Final Demo, One More Time

See the Pen Faded Outer Box Backdrop Blur [forked] by Yair Even Or.

I’ve also prepared a simplified version with minimal code and no CSS variables that’s easier to read and re-purpose.

20 Dropbox Tips and Tricks for 2024 (Updated)

Original Source: https://www.hongkiat.com/blog/dropbox-tips-and-tricks/

If you think of Dropbox as just another cloud storage service, you might be surprised by the depth of its features. Beyond basic file saving and sharing, Dropbox offers a plethora of lesser-known functionalities that enhance its utility and performance.

Known for its simplicity and power, Dropbox stands out among cloud storage options. This post will unveil various hidden tips and tricks to help you master Dropbox, enabling you to do much more than merely store and sync files. Let’s dive into these insights.

Table of Content

Request files, get in Dropbox
Preview any file format
Protect Dropbox account with two-step verification
Access deleted folders
Track everything you’ve shared
Check recent activities of your files
Collaborate on files with teams
Customize Dropbox notifications
Access files without the internet
Boost sync speed
Delete files without traces
Access folder versions of files
Smart bookmarks with Dropbox
Get direct download links
Share screenshots quickly
Add comments to files
Backup desktop and document folders
Custom sync files and folders
Check Dropbox file activities
Auto-sync memories

Request files from anyone, get them in Dropbox

Did you know that anyone can send files to your Dropbox, even if they don’t have an account? The File Requests feature makes this possible. It allows you to collect and receive files of any size directly into your Dropbox, offering a convenient way for others to send you what you need.

Here’s a step-by-step guide to request files anonymously:

Log in to your Dropbox account.
Access the “File requests” option from the left sidebar.
File requests section in DropboxFile requests section in Dropbox
Select “Create a file request”.
Provide a brief description for your request and choose “Change folder” to specify where the incoming files should be stored. Then, click “Next”.
Creating a file request in DropboxCreating a file request in Dropbox
Copy the provided link and share it with anyone from whom you wish to receive files. Alternatively, enter their email addresses and hit “Send” to invite them directly.
Sending a file request in DropboxSending a file request in Dropbox

Preview almost any file format

Struggling to open a file on your computer? Upload it to Dropbox and use its versatile “file preview” feature. This handy tool lets you peek into various file types such as PDF, PSD, AI, PNG, JPG, SVG, and BMP, among others.

Previewing a Photoshop file in DropboxPreviewing a Photoshop file in Dropbox

Enhance your productivity with the integration of Office Online, allowing you to read and edit documents, spreadsheets, and presentations right from your browser for free.

Here’s how to preview or edit files in Dropbox:

Click on any file to open it in Dropbox.
In the preview window, look for the arrow next to the “Open” button and click it.
Select from options like “Microsoft Word/Excel/PowerPoint Online” to start editing.
Editing a document in DropboxEditing a document in Dropbox

Protect your account with two-step verification

To enhance the security of your Dropbox account against unauthorized access, it’s highly recommended to enable two-step verification. This security measure introduces an additional layer of protection by sending a unique code to your phone each time a sign-in is attempted from an unfamiliar device.

Though it may seem like an additional task, the increased security for your account is substantial. Activating this feature is straightforward and can be done through the security settings of your account. Here’s how:

Sign in to dropbox.com using your credentials.
Navigate to Settings > Security.
Security Settings
In the “Two-step verification” section, choose and set up your preferred verification method.

Re-access deleted or left-shared folder

Ever left or deleted a shared folder and needed it back? With Dropbox, retrieving it is straightforward. Here’s how you can re-access shared folders:

Go to “Sharing” in the left sidebar of your Dropbox.
Under the “Folders you can add to your Dropbox” section, find the folder you wish to rejoin and click “Add”.
Adding a previously shared folder in DropboxAdding a previously shared folder in Dropbox

Track everything you’re sharing

Ever wondered what exactly you’ve shared on Dropbox? It’s easy to keep track and ensure you haven’t accidentally shared something unintended. Dropbox provides a straightforward feature to view all your shared items.

Simply log in to your Dropbox account and select the ‘Shared’ option in the left sidebar.

You’ll see a list of all your shared folders and files, including last modification dates. This list encompasses both the folders you’ve directly shared and those you have access to but haven’t added to your space. You can also manage who has access, adding or removing contributors as needed.

Shared Files and Folders

For a detailed look at each link-shared file or folder, click on the ‘Links’ tab. Here, you can review the sharing date for each item and, if necessary, delete any link to revoke access, ensuring complete control over your shared content.

Check recent activities of your files

Looking for the latest updates or the last file you worked on? Dropbox makes it easy with the “Recents” section in the left sidebar. This feature provides an updated list of recently opened or modified files, so you’re always in the loop.

From here, you can share, download, comment, delete, or view previous versions of your files with ease.

Checking recent activities in DropboxChecking recent activities in Dropbox

Work on files as a team

Team collaboration is a breeze with Team Dropbox. It’s a feature designed for groups to share and manage files efficiently, offering different privileges to team members. And the best part? It’s free!

Start by sharing a single folder with the entire team or specific groups. You can manage access, organize content, and even create groups to share files with everyone in the group instantly, similar to Friend Lists in Facebook.

Setting up a new Team Dropbox account is easy:

Select the “Team” option from the left sidebar in Dropbox and choose “Create free team”.
Enter your team’s name, create a new Team Dropbox, and provide your personal/work email and password for the team, then click “Continue”.
Creating a new team in DropboxCreating a new team in Dropbox
Invite members to your team as needed and proceed by clicking “Continue”.
Inviting members to Dropbox teamInviting members to Dropbox team
Finalize by clicking “Create team Dropbox”, and you’re all set.

Customize Dropbox notifications

For those heavily using Dropbox, the influx of email and push notifications can become overwhelming. Fortunately, customizing these alerts is a straightforward process to keep only the most relevant updates coming your way.

Managing Email Notifications

Begin with email notifications by heading to Dropbox.com and adjusting the ‘Notification settings’ under your ‘Personal account’. Tailor these settings to receive only the notifications you find most pertinent.

Email Notification Settings
Configuring Desktop Notifications

For desktop alerts, dive into your Dropbox settings and select the ‘Notifications’ tab under Preferences. This area allows you to specify which updates you want to see on your desktop.

Desktop Notification Settings

By fine-tuning these settings, you can streamline your notifications to ensure you’re only alerted about the most critical updates, helping maintain focus and reduce unnecessary interruptions.

Access files without Internet

Need to access certain files on the move, but stuck without internet? No problem. Dropbox’s Offline feature allows you to keep important files accessible on your mobile device, anytime, anywhere.

To make files available offline, simply:

Open the Dropbox mobile app and navigate to the files you want to access offline.
Tap the icon to the right of the file and switch on the “Available offline” toggle. Dropbox will download the files to your device, making them accessible without an internet connection.
Access these offline files anytime by tapping on “Offline” in the app.
Accessing offline files in Dropbox on mobileAccessing offline files in Dropbox on mobile

Boost synchronization speed

Dropbox is designed to be considerate of your internet usage, typically using only 75% of your maximum upload speed. However, if you’re looking to speed up your uploads or syncs, you can adjust this setting.

To maximize Dropbox’s bandwidth usage:

Right-click the Dropbox icon in your system tray.
Click the gear icon and select “Preferences”.
Accessing Dropbox preferencesAccessing Dropbox preferences
Go to the “Bandwidth” tab.
Adjusting bandwidth settings in DropboxAdjusting bandwidth settings in Dropbox
Choose “Don’t limit” under the “Upload rate” to allow Dropbox to use the full speed of your internet connection for uploads.

Delete files without leaving tracks

Accidentally stored a sensitive file in Dropbox and need to remove it discreetly? Dropbox’s cache folder, “.dropbox.cache”, is a hidden folder where deleted files are temporarily stored. Clearing this cache ensures that files are completely removed.

To clear the Dropbox cache and delete files permanently:

Open your file manager and enter %HOMEPATH%Dropbox.dropbox.cache in the location bar, then press Enter.
This takes you directly to the Dropbox cache folder, where you can permanently delete any files or folders you no longer want.
Navigating to Dropbox's cache folderNavigating to Dropbox's cache folder

Access older versions of files

Ever wish you could undo changes to a file? With Dropbox, you can! It keeps older versions of your files for 30 days, allowing you to revert to previous versions if needed, much like the File History feature in Windows 10.

Here’s how you can access an older version of a file:

Next to the file of interest, click the three dots and choose “Version history”.
Accessing version history in DropboxAccessing version history in Dropbox
Preview the versions listed and click “Restore” on the one you want to revert to.

Restoring a previous version in DropboxRestoring a previous version in Dropbox

Smart bookmarks with Dropbox

Imagine having all your bookmarks accessible from any device via the cloud. Dropbox introduces a seamless way to store and sync your bookmarks just by dropping them into a designated folder.

Note: While this works great with Google Chrome and Mozilla Firefox, Microsoft Edge does not support the drag-and-drop feature for this purpose.

To organize and sync your bookmarks using Dropbox:

Create a “Bookmarks” folder in your Dropbox. This will be your central hub for all links.
Creating a “Bookmarks” folder in DropboxCreating a “Bookmarks” folder in Dropbox
Simply drag the small icon next to the browser’s location bar and drop it into your “Bookmarks” folder in Dropbox.
Dragging a bookmark into DropboxDragging a bookmark into Dropbox

Get direct download links

Typically, when you click a Dropbox link, it opens a preview page with an option to download. But there’s a neat trick to bypass this and start the download immediately.

For a direct download link, change the dl=0 at the end of the Dropbox link to dl=1. This simple tweak turns it into an instant download link.

Typical Dropbox link with “dl=0”Typical Dropbox link with “dl=0”
Modified Dropbox link with “dl=1” for direct downloadModified Dropbox link with “dl=1” for direct download

Share screenshots on the fly

Gone are the days of manually dragging screenshots to Dropbox to share them. With Dropbox’s screenshot sharing feature, you can streamline the process and share your screenshots instantly.

To set up quick screenshot sharing:

Right-click the Dropbox icon in your system tray.
Click the gear icon and choose “Preferences”.
Accessing Preferences in DropboxAccessing Preferences in Dropbox
Navigate to the Import tab, select “Share screenshots using Dropbox”, and click “OK”.
Enabling screenshot sharing in DropboxEnabling screenshot sharing in Dropbox
Now, when you take screenshots using the Ctrl + Print Screen keys, Dropbox will automatically save them to a “Dropbox/screenshots” folder and copy the link to your clipboard, ready for quick sharing.

Add comments to files

Dropbox not only makes file sharing effortless but also includes a powerful commenting feature to facilitate discussions directly on any file. Whether you’re inviting someone to view a file or sharing a link, don’t miss out on this collaborative tool.

Accessible via Dropbox’s web, mobile, and desktop applications (specifically the Dropbox app, not the file browser on your computer), this feature allows you to select any file or view it in preview mode to access the comment sidebar.

In this sidebar, you can input your observations, ask questions, or tag teammates with an @ to bring them into the conversation. They’ll receive a notification either in Dropbox or via email, allowing you to see the dialogue unfold in real time.

This commenting tool centralizes all discussions related to a file, ensuring that every detail is noted and nothing is missed. As you and your team update the file, the online preview updates as well, keeping everyone in sync with the latest version.

Back up Desktop and Documents folders

Expand your Dropbox’s utility beyond just syncing a specific folder—now it can also protect the entirety of your Desktop, Downloads, and Documents folders.

Backup Desktop and Documents

Setting up is straightforward:

Access your Dropbox settings,
locate the ‘Backups’ section, and
select the ‘Manage backups’ option.
Manage Backups

Choose the folders you wish to sync and back up.

This functionality offers a handy way to safeguard essential files outside your main Dropbox folder, ensuring your Desktop and other crucial areas are always up-to-date and secure in Dropbox with minimal effort.

Sync required files and folders only

If you’re looking to save space on your computer or conserve bandwidth, Dropbox’s Selective Sync feature is your ally. It allows you to choose only the folders you need to sync, keeping your hard drive clutter-free and your Internet usage optimized.

To enable Selective Sync:

Right-click the Dropbox icon in your system tray.
Click the gear icon and select “Preferences”.
Selecting Preferences for DropboxSelecting Preferences for Dropbox
Go to the “Account” tab and click “Selective Sync”, then hit “OK”.
Account tab for Selective Sync in DropboxAccount tab for Selective Sync in Dropbox
In the Selective Sync window, deselect any folders you do not want to sync to your computer. Click “Update”, then “OK” to confirm your changes.
Configuring Selective Sync in DropboxConfiguring Selective Sync in Dropbox

Note: Deselecting folders will remove them from your computer’s hard drive but keep them accessible through dropbox.com and other devices.

Check Dropbox files activities

The “Events” feature in Dropbox is a powerful tool that logs various activities within your account, such as file creation, sharing, deletion, and restoration. This log is not just a timeline but a detailed record that you can filter by category and date, and sort for easier navigation.

To access the event log in Dropbox:

Click on “Events” in the left sidebar of your Dropbox page.
Explore the timeline to see a detailed log of all activities associated with your account.
Exploring the event log in DropboxExploring the event log in Dropbox

Auto-sync your memories

Dropbox isn’t just for files and folders; it’s also an ideal space for storing and viewing your photos and videos. With its online gallery feature, you can easily view your media without the need to download them, making Dropbox a perfect place to safeguard and enjoy your precious memories.

To view your photos and videos in Dropbox:

First, upload your photos and videos to Dropbox.
Go to dropbox.com and select “Photos” from the sidebar to access your media.
Browse through your photos and videos, neatly organized in a gallery format.
Browsing the photo gallery in DropboxBrowsing the photo gallery in Dropbox

Bonus: Sync data of portable apps

Dropbox’s synchronization isn’t limited to documents and media; it can also keep your application data consistent across multiple computers. This feature is particularly handy for portable apps, allowing you to have the same settings, bookmarks, and history across all devices without needing to rely on individual app sync features.

Here’s a guide to syncing a portable app’s data using Dropbox, demonstrated with Thunderbird:

Download and install the portable version of an app (e.g., Thunderbird) directly into your Dropbox folder.
Installing a portable app in the Dropbox folderInstalling a portable app in the Dropbox folder
Run the portable app from the Dropbox folder as you would normally. Any changes or data will be automatically synchronized across all your devices through Dropbox.

Explore these lesser-known tips to maximize your productivity and make the most out of your Dropbox experience. Happy dropboxing!

The post 20 Dropbox Tips and Tricks for 2024 (Updated) appeared first on Hongkiat.

Bold Typography Shines in HOMME's Branding & Visual Identity

Original Source: https://abduzeedo.com/bold-typography-shines-hommes-branding-visual-identity

Bold Typography Shines in HOMME’s Branding & Visual Identity
Bold Typography Shines in HOMME's Branding & Visual Identity

abduzeedo0122—24

Discover HOMME’s new branding and visual identity, where organic typography and bold fonts merge to create a standout design. Explore this innovative project on abduzeedo.com

In the diverse and ever-evolving world of design, every project brings with it a unique set of challenges and opportunities. The recent branding and visual identity project for HOMME, created by Charlotte Slegers, stands out as a prime example of innovative design. This project, as seen on Behance, skillfully combines organic typography elements with a daringly bold font, creating a distinct visual language that speaks volumes about the brand’s character.

Charlotte Slegers, a notable figure in the design community, has applied her expertise to develop a visual identity that is both refreshing and impactful. The project hinges on a clever representation of the double “m” in HOMME, which adds a layer of personality and distinctiveness to the brand. This design choice is not just a visual gimmick; it’s a strategic move that encapsulates the essence of the brand in a single, memorable feature.

What sets this project apart is how the typography is seamlessly integrated across various mediums. From the website to print media, the consistency and fluidity of the design ensure a cohesive brand experience. This kind of integration is crucial in branding, as it reinforces brand recognition and builds a strong visual connection with the audience.

The choice of a super bold font aligns perfectly with the current trends in typography. Bold fonts are not just visually striking; they convey strength and confidence, qualities that are desirable in any brand. This approach also demonstrates a keen understanding of how typography can influence perception and evoke certain emotions in the viewer.

The visual identity of HOMME, as curated by Charlotte Slegers, is more than just a branding exercise. It’s a testament to the power of typography in shaping a brand’s image. The project showcases how a well-thought-out typographic strategy can elevate a brand’s identity, making it stand out in a crowded marketplace.

This project is a must-see for anyone interested in the intersection of typography and branding. It serves as an inspiration and a benchmark for designers looking to create compelling and effective visual identities.

Branding and visual identity artifacts

Branding Brand Design brand identity identity logo Logo Design logos Logotype visual identityBranding Brand Design brand identity identity logo Logo Design logos Logotype visual identityBranding Brand Design brand identity identity logo Logo Design logos Logotype visual identityBranding Brand Design brand identity identity logo Logo Design logos Logotype visual identityBranding Brand Design brand identity identity logo Logo Design logos Logotype visual identityBranding Brand Design brand identity identity logo Logo Design logos Logotype visual identityBranding Brand Design brand identity identity logo Logo Design logos Logotype visual identity

For more information make sure to check out Charlotte Slegers website and Behance profile.