How a molecular structure explains the transport of fatty acids past the blood-brain barrier

The brain and eyes develop through constant circulation of nutrients through the blood-brain and blood-retina barriers. One such nutrient that is essential for development is an omega-3 fatty acid called docosahexaenoic acid (DHA). DHA makes up a fifth of all the fatty acids required on the membranes of cells in the central nervous system. Neither the neurons in the brain nor the cells in the eye are capable of synthesizing DHA by themselves and therefore depend on dietary sources for DNA. Previously, scientists knew from cellular clues that this fatty acid most likely passed through to the blood-brain and blood-retina barriers in the form of lysophophatidylcholine (LPC-DHA) using a molecular channel. This transporter is known as a major facilitator superfamily domain containing 2A, or MFSD2A, with the help of sodium atoms regulating the channel. However, it was not clear how this channel allowed the passing of complex molecules like DHA. A recent study by Dr. Rosemary Cater and colleagues at Columbia University provided precise clues to further show the structure of this channel. 

To investigate the structure of MFSD2A, the authors used a state of the art imaging technique called single-particle cryo-electron microscopy. This is a method of electron microscopy where a beam of electrons is transmitted through a rapidly frozen purified molecule. Because the sample is flash frozen, the molecules trapped in a frozen state can be imaged in their native shape as present in the cell and from multiple angles. By capturing and combining multiple captured 2D images, a 3D structure of the protein can be reconstructed with extreme accuracy. Cryo-electron microscopy is so impactful in biological significance that this method was awarded the 2017 Nobel Prize in Chemistry. found a number of molecular patterns and arrangements of protein chains that make up a full molecule of MFSD2A

Protein structure studies are typically among the most challenging grounds to explore in biology because proteins need to be captured in their native state as present in the cell. Past discoveries of various protein structures have been so instrumental in shaping therapeutic areas that the extent of mechanistic understanding of biological molecules has resulted in recognition by Nobel committees. Most recently, the discovery of the structure of the ribosome opened up fields of exploration into therapeutic interventions into ribosome diseases, some of which can lead to cancer.

To get the best chance at imaging the structure of MFSD2A, the scientists extracted and examined purified versions of this protein obtained from multiple organisms: the zebrafish, the frog, European cattle, the domestic dog, the red junglefowl, mice and humans. Finally, the authors found that the protein obtained from the red junglefowl, which is a rooster species that originates from Southeast Asia, was the most experimentally stable and most alike (73% identity) the human version of MFSD2A. 

Using additional accessory proteins to help with the orientation of MFSD2A, the authors obtained high-quality images, with a resolution of 0.3 nanometer, or 0.3 billionth of a meter. From the imaging data, the authors found that MFSD2A protein itself is about 5 nm wide and 8 nm long. MFSD2A is a transporter protein and like many transporters, it contains repeated bundles of helices made of protein chains that traverse the cell membrane and are connected by a protein chain that loops within the space in the cell. 

Structure of MFSD2A arranged as protein helices (colored cylinders) within the cell membrane along with protein loops that form both in the extracellular space (“Out”) and within the interior, cytoplasmic space (“In”) of the cell. The cytoplasmic loops likely have an important functional role. Figure from Cater et al, 2021.

The cell membrane consists of two layers of lipid molecules, known as the lipid bilayer, that allow entry and exit of materials from the cell. These loops provide the shape to the protein inside the cell such that it appears to provide a large enough cavity opening from the lipid bilayer into the cellular space to allow the target molecules to enter the cell. Amino acids are the building blocks of proteins and the cavity contains amino acids of both water-attracting and water-repelling kind. This property makes it possible for many molecules of differing chemical nature to be able to be accommodated within the cavity. This cavity contains three important regions that allow for the protein to be specific and functional: a charged region, a binding site for sodium atoms and a lipid-specific pocket. The authors speculate that these parts help in establishing the mechanism by which LPC-DHA is transported from the outside into the cell. The multiple protein helices form two protein domains that capture LPC-DHA from outside the cell layer of the blood brain barrier of endothelial cells, then rock over a rotation axis so that now their confirmation switches and finally, they release the protein molecule into the cell. For this activity of movement of LPC-DHA, sodium atoms are absolutely required to allow for the shape change of the protein. Once LPC-DHA enters the barrier cells in this manner, the protein is then transported across to the other side of the cell facing the brain containing neurons. 

The transporter channel MFSD2A changes its shape once it binds sodium atoms in the extracellular space, which helps the transport of LPC-DHA from the blood into the brain space through the barrier of a single line of cells made up of endothelial cells. Figure adapted from Cater et al, 2021.

Humans with mutations in MFSD2A gene have abnormal brain defects such as microcephaly, and disruption of the gene in mice affected neuronal branching and fatty acid composition in the brain. The discovery of the structure of a molecule that mediates uptake of essential nutrients across the blood-brain and eye-brain barriers will help in the delivery of therapies of neurological diseases.

Dr. Rosemary J. Cater is a postdoctoral researcher in the lab of Dr. Filippo Mancia in the Department of Physiology and Cellular Biophysics at Columbia University.

Take a Break: How the Brain Chooses When to Explore and When to Rest

Have you ever wondered why we feel comfortable in a familiar place or why going back to our favorite spots over and over again feels so good? Well, Dr. Paolo Botta, a former postdoc at Columbia University, and colleagues attempted to unravel some of the inner workings of the brain when it comes to rest and exploration. More specifically, Dr. Botta examined how neuronal activity correlates with periods of rest when exploring new areas. Dr. Botta and colleagues followed the behavior of mice as they freely explored a new area. They specifically looked at where and how often these mice decided to exhibit arrest behavior, or, in other words, take a break during their explorations. While the arrest behavior alone is a fascinating phenomenon and provides insight into how mice explore new spaces, Dr. Botta and colleagues decided to go a step further and see which neurons in the brain are important for this arrest behavior. They decide to home in on an area of the brain called the Nucleus of the Basal Lateral Amygdala (BLA). This area has previously been shown to be involved in locomotor exploration, experience based learning, recognition of familiar areas.

With this information in hand, Dr. Botta and colleagues began by identifying whether BLA neurons are active during arrest behavior. To this end, they gave mice access to both their home cages and a large open area for five days and allowed them to freely explore the large open area during this period. BLA neuronal activity was monitored in the mice by measuring calcium levels, with higher calcium levels indicating neuronal activity (Figure). The researchers observed an increase in calcium in BLA neurons during arrest behavior, which means that BLA neurons are involved in this type of behavior.  However, do these neurons actually cause the arrest behavior? To answer this question, Dr. Botta and colleagues either activated or inhibited the neurons using optogenetics. Optogenetics is a technique in which neurons are stimulated by light. So, by turning different lights on and off, the researchers were able to either activate or inhibit BLA neurons whenever they wanted to. When they activated the BLA neurons, the mice decreased their speed and experienced more arrest behavior. When they inhibited the BLA neurons, the mice had an increase in movement speed. After seeing how turning BLA neurons on and off affected behavior, they concluded that the BLA neurons are important for inducing arrest behavior.

At this point, Dr. Botta and colleagues have revealed that BLA neuronal activity occurs specifically during these arrest behaviors and that their activity is important for the onset of the arrest. However, their curiosity did not stop there. They began to wonder whether BLA activity changed when the mice exhibited arrest behavior, or took breaks in more familiar areas. To figure this out, Dr. Botto and colleagues tracked exactly where the mice explored and counted how many times the mice exhibited arrest behavior in areas that they previously explored. With this experiment, they realized that the mice were more likely to exhibit arrest behavior in areas previously visited. So, mice, like humans, have favorite spots and they like to rest in those spots! After seeing that the mice have favorite spots, Dr. Botto and colleagues went on to examine the BLA neuronal activity in these familiar areas. They found that there was an increase in neuronal activity in these familiar areas and the more a mouse revisited and exhibited arrest behavior in a specific area the more neuronal activity developed. In other words, the more often a mouse took a break in a specific area the more that correlated with BLA neuronal activity.

The amygdala has multiple nuclei, which consist of groups of cells that are important for specific roles. The Central Nucleus of the Amygdala (CEA) is a part of the amygdala that has previously been shown to be involved in immobility. BLA neurons also communicate with the CEA (Figure). Knowing that the BLA neurons are important for invoking arrest behavior and the CEA plays a role in immobility, Dr. Botta and colleagues were curious as to whether these BLA neurons that project to the CEA are the specific neurons involved in triggering arrest behavior. To see whether the BLA neurons that project to the CEA are the ones active during arrest behavior they used the combination of calcium imaging and optogenetic techniques previously mentioned. With these techniques they were able to see that the BLA neurons that project to the CEA had an increase in neuronal activity during arrest behavior (Figure). This increase was not seen in BLA neurons that projected to other parts of the amygdala indicating that the BLA-CEA interaction is integral for the arrest activity. They also repeated the stimulation of the BLA neurons that project to CEA and observed an increase in arrest while inhibiting the same neurons resulted in an increase in movement, further confirming the need of this BLA-CEA interaction to induce arrest behavior.

Overall, Dr. Botto and colleagues discovered that BLA neurons that communicate with the CEA are important for arrest behavior, particularly in familiar places. This behavior seems to be extremely important for allowing a mouse to orient itself and properly explore novel surroundings. Maybe humans have a similar pathway that we use when wandering around. Could my BLA be the reason why I always go to the same cafes after a long walk or stop in the same part of the park while walking my dog? Are our BLA neurons just firing away while we rest? 


Figure: BLA neuronal activity during exploratory vs arrest behavior.      Left: Decreased activity in BLA neurons that communicate with the CEA results in increased exploratory behavior. Right: Increased BLA to CEA neuronal activity, indicated by calcium signaling, results in increased arrest behavior. Red colors indicate decreased BLA neuronal activity and increased exploratory behavior. Green colors indicate increased BLA neuronal activity and increased arrest behavior. BLA: Nucleus of the Basolateral Amygdala, CEA: Central Nucleus of the Amygdala

Why the gallbladder matters – The role of bile acids in metabolic health

Unless you belong to the 10-15% of people that have gallstones, you probably never think about your gallbladder or its function. However, this small pear-shaped organ plays an important role in our digestive system. The gallbladder is situated right under the liver, and stores bile produced by liver cells. Mostly after eating meals, bile is released from the gallbladder into the gut. Here, substances within the bile called bile acids help with the breakdown and absorption of fat. Apart from their role in the digestive system, bile acids have been shown to communicate with other organs and thereby affect the metabolism of fat and sugar. Not unexpectedly, considering their roles in digestion and metabolism, bile acids are associated with various metabolic diseases in humans, such as obesity and diabetes. Modulation of bile acids could therefore be used as a strategy to treat or prevent such metabolic disorders.

Dr. Antwi-Boasiako Oteng and colleagues from the Haeusler lab of the Department of Pathology & Cell Biology at Columbia University aimed to better understand how bile acids influence metabolic health. The two primary bile acids in humans are cholic acid (CA) and chenodeoxycholic acid (CDCA), which are both made from cholesterol in the liver (see Figure below, left panel). CA production requires the enzyme Cyp8b1, which adds a hydroxyl group (i.e., one hydrogen atom bonded to one oxygen atom) to the 12th carbon of the molecule. Because of this modification, CA is called a “12α-hydroxylated bile acid”. CDCA does not contain this hydroxyl group on its 12th carbon, and is therefore called a “non-12α-hydroxylated bile acid”. Levels of 12α-hydroxylated bile acids like CA are higher in obese individuals, but it is not yet known whether they can cause obesity. Therefore, in their study published in Molecular Metabolism, the researchers investigated the role of non-12α-hydroxylated versus 12α-hydroxylated bile acids in the development of metabolic disorders in an experimental mouse model.

Dr. Oteng and colleagues started their research by genetically manipulating mice in such a way that they have bile acids more similar to humans. This is necessary as mice produce bile acids called muricholic acids (MCAs) that are not present in humans. By removing the enzyme that converts CDCA into MCA, the researchers created mice with a human-like bile acid profile (see Figure below, right panel). Aside from having low levels of MCA, these mice had higher levels of CDCA and lower levels of CA as compared to regular mice. Most importantly, fat absorption was strongly reduced in these mice, and they were protected from weight gain when fed a high-fat diet. To test if this could be due to the relative decrease in 12α-hydroxylated bile acids, the researchers supplemented another group of human-like mice with high levels of 12α-hydroxylated bile acids. This treatment strongly promoted fat absorption, which may increase susceptibility to metabolic disorders when combined with an excessive and/or unhealthy diet.

These findings suggest that the ratio of non-12α-hydroxylated versus 12α-hydroxylated bile acids is an important determinant of metabolic health in humans, which opens up new avenues for therapeutic intervention. According to Dr. Oteng, “We now have increased confidence that targeting Cyp8b1 to reduce the levels of 12α-hydroxylated bile acids can reduce the risk of metabolic disease in humans”. Currently, more than one-third of Americans have metabolic disorders, which increases their risk of heart disease and other health problems. If proven effective, a therapy targeting human bile acid composition could have a major impact on public health.


Left panel. The bile acids CA and CDCA are made from cholesterol in the liver. CA is 12α-hydroxylated bile acid produced by the enzyme Cyp8b1, while CDCA is non-12α-hydroxylated. In mice, CDCA is further converted into MCA. Right panel. Inhibition of the conversion of CDCA into MCA increases the ratio of non-12α-hydroxylated versus 12α-hydroxylated bile acids, thereby reducing fat absorption in the gut and protecting from diet-induced weight gain.

The key to a longer life might be in skipping that midnight snack

Have you ever caved in to the temptation of a snack in the middle of the night that manifested into a quick freezer dive to grab that ice-cream or into a series of quick taps on your food delivery app to get those udon noodles? Suffice it to say that I have been a victim to this thought one too many times. Much to my chagrin, there is an abundance of evidence that suggests that eating during restricted hours of the day or time-restricted feeding (TRF) can slow down decline of bodily functions. Limiting food intake to certain hours of daytime, even if the food is not necessarily nutritious or low in calories, can prevent ageing or even kickstart anti-aging mechanisms in mice and flies with obesity or heart disease. Because ageing was dependent on when the body takes in food, these studies hint at the role of the body’s biological clock, known as circadian rhythms, in regulating health and longevity. In an unexpected new study authored by Columbia postdoc Dr. Matt Ulgherait, flies following time-restricted feeding while also balancing it with an unlimited all-access ad libitum diet, show a significant increase in lifespan. 

By structuring 24 hour day-night periods as cycles of 12 hours of light followed by 12 hours of darkness in a temperature-controlled box, the authors tested various dietary regimens for their effects on lifespan and stumbled upon one regimen that consistently showed longer lifespan along with enhanced health in the flies. This regimen cycled between a 20-hour fast starting at mid-morning (6 hours after lights on) to a 22 hour recovery period of eating ad-libitum on repeat in young flies within 10-40 day post hatching stage of adulthood. However, flies that began this regimen after reaching older age at day 40 did not show enhanced lifespan. In comparison to flies that were allowed access to food ad-libitum on a 24 hour cycle, flies following this particular fasting-feeding regimen showed a 18% increase in female lifespan and 13% increase male lifespan in their young age. Due to the cycling schedule of unlimited food access with periods of fasting, the authors termed this regimen as intermediate time-restricted feeding (iTRF).

Previous studies have shown that caloric restriction through reduced food intake, protein restriction or inhibiting insulin-like signaling can extend lifespan. However, iTRF did not appear to limit flies from eating less and in many cases, resulted in flies eating more during times of food access compared to those in the ad libitum group. Thus, lifespan extension under iTRF did not occur because of limitation in nutrient uptake. Interestingly, an iTRF regimen performed under additional treatments of either dietary protein restriction or inhibited insulin-like signaling, resulted in a marked boost in lifespan compared to iTRF alone. It therefore seems that  independent mechanisms that  can enhance lifespan can be combined to increase lifespan even more. 

While these methods provide ways to extend lifespan through incremental means, some might argue that it would be meaningless to simply survive without long-lasting health benefits. To examine whether the longer-lived flies continued to exhibit youth, scientists measured the fitness of the flies using two well-known age-related tests: the flies’ ability to climb up the plastic vial they are in and how much they accumulate in their tissues aggregates of aging proteins – polyubiquitin and p62. When compared to the ad libitum group, iTRF flies climbed much faster and had fewer polyubiquitin and p62 aggregates in the flight muscles, even after they reached an age beyond 40 days of hatching. While the gut microbiome was shown to dictate proclivity for disease and thus have an effect on lifespan, the gut tissue in iTRF flies remained healthier with more normal cells, even when the gut microbiome was depleted with antibiotics. Therefore, the flies appeared to be in optimal health conditions with fewer aging markers in addition to longer survival, demonstrating yet again that aging slowed down due to better functioning of organs.

The dietary regimen under iTRF only controls the timing of feeding but not the nutritional intake, which provided clues to the authors that perhaps the body’s natural biological clock had something to do with iTRF-mediated lifespan. The biological clock in flies consists of proteins that are also present in other organisms all the way from fungi to humans. The main molecular parts of the core circadian clock include the proteins ‘Clock’ (Clk) and ‘Cycle’ (Cyc) which activate the genes period (per) and timeless (tim), which in turn inhibit Clk and Cyc. This process is called a feedback loop which takes all of 24 hours to complete in both flies and humans, and this is how our bodies respond to light-dark cycles. Flies undergoing iTRF showed enhanced expression of Clk in the daytime and of per and tim at night time. The authors then explored the feeding behavior and metabolism of circadian clock genetic mutants undergoing iTRF and found that neither the 20 hour long fasting period nor dietary restriction in their food altered their feeding behavior when compared to normal flies under iTRF. Yet, the extended lifespan was completely missing in Clk, per and tim mutants undergoing iTRF. Even the improved health seen with an iTRF regimen through better climbing ability and less aging-protein aggregation was abolished in per mutants compared to normal flies. Shifting the iTRF cycle by 12 hours with a fasting period during the daytime abolished the occurrence of an extended lifespan. In the altered regimen, while the same cycle was now only shifted by half a day, eating at night time while fasting during the day just did not work. This discovery showed that there could be a deep link between the body’s biological timer and when during the day food is eaten that determines both longevity and well-being. 

Because shifting the fasting period to daytime did not show any benefits, the authors checked whether genes that activate during fasting are also linked to the biological clock. In fact, Dr. Ulgherait and group had already shown that disrupting tim and per genes in the gut, which is where food is processed, caused an increase in lifespan. But, iTRF included periods of starvation that could trigger different metabolic processes. Starvation induces cellular mechanisms to degrade and recycle its molecules in a process called autophagy. Interestingly, genes encoding two autophagy proteins, Atg1 and Atg8a, which are also present in humans, showed peak levels in the night time with enhanced peaks in flies under iTRF. During autophagy, there is an increased activity of cell organelles called lysosomes that contain digestive enzymes needed to break down cellular parts. The authors found that normal flies fasting under iTRF showed higher Atg1 and Atg8a expression along with more lysosomal activity but period mutants failed to do so. Using some more genetic tricks, the authors found that manipulating the level of autophagy to go up or down directly showed an effect on iTRF-mediated lifespan.

Finally, to explore the link between iTRF-mediated lifespan and autophagy, the authors used genetic tools to increase night-specific levels of Atg1 and Atg8a. In a surprising revelation, flies with night-specific expression of Atg1 and Atg8a showed an increase in lifespan, even when these flies did not undergo fasting and were fed ad libitum. Subjecting these genetically altered flies to iTRF did not additionally increase their lifespan, suggesting to the authors that circadian enhancement of cellular degradation under an all-access diet provides the same beneficial effects as fasting done under the stricter regimen of iTRF. Flies with night-specific enhanced autophagy also showed better neuromuscular and gut health on an all-access diet. Therefore, clock-dependent enhancement of the biological recycling machinery can mimic the lifespan extension mediated by iTRF.

 

Now of course large genetic manipulations are not yet a consideration in humans but this study provides a potentially powerful yet simple change in dietary strategy that could just somehow slow down aging. Aging increases risk of mortality and disease but imagine a food intake regimen translatable from this study into humans that can help improve overall neuromuscular and gut health. So, while technology has indeed made it so much easier than before to have food at our doorstep in a few phone taps in the middle of the night, perhaps restricting the hours of when we eat can really help us live healthier lives. This study now makes me reconsider the famous quote by Woody Allen in the context of food, “You can live to be a hundred if you give up all the things that make you want to live to be a hundred”.

Dr. Matt Ulgherait is a postdoctoral researcher in the lab of Dr. Mimi Shirasu-Hiza in the department of Genetics & Development at Columbia University. Dr. Ulgherait and his colleagues also recently showed that removing the expression of the period gene from the gut tissue was sufficient to cause an increase in lifespan.

Magic under the microscope

Researchers design an accessible, straightforward technique to characterize moiré systems – a class of materials built by placing slightly misaligned atomic monolayers on top of each other. Under certain conditions, such moiré structures exhibit exotic physical phenomena absent in the individual units that conform them.

A moiré pattern is an interference effect that arises when two grids are superimposed. It can be observed in the wrinkles of a mesh shirt and it is responsible for the fringes that appear when taking a picture of a computer screen. Moiré patterns are present in art and fashion, and in the last few years their effect in two-dimensional materials has entailed a revolution in physics.

Two-dimensional materials are those that are less than a nanometer thick. The first one to be isolated was graphene, a single-layer of carbon atoms (see Fig. 1a). Such discovery opened a whole new field of research and many labs around the world started making their own stacks – structures with two-dimensional materials placed on top of each other. If one were to place one of those layers slightly misaligned with the one below, a moiré pattern would emerge. This interference effect can be visualized in Fig. 1b. The small circles represent the carbon atoms that form a crystalline lattice (an ordered structure) on each graphene layer. The top layer is rotated with respect to the bottom one and, as a consequence, a periodicity larger than the atomic lattice emerges as highlighted in Fig. 1b.

In 2018, the field of condensed matter physics was stirred up: such moiré materials, at a very specific misalignment value called the magic angle, exhibit electronic states of matter that are not present in the individual layers, such as superconductivity or magnetism. The emergence of those electronic phases is a consequence of the moiré pattern and its direct visualization is critical for their understanding. There are a few techniques, including transmission electron microscopy and scanning tunneling microscopy that allow for this, but they require complex setups that do not necessarily work for any material, which has significantly slowed down the progress in the field. McGilly and colleagues show a new and simple technique based on piezoresponse force microscopy to visualize moiré patterns.

A piezoresponse force microscope consists of a sharp metallic tip brought into contact with the material under study –  in this case, the moiré system (see Fig. 1c). Piezoresponsive materials are those that undergo a mechanical deformation in the presence of an electric field. In the microscope, the sample moves a small amount when a voltage is applied across it and the tip follows the motion. Such tip motion is measured as a voltage which is amplified to detectable values. The tip is then moved around the sample and the process is repeated on every pixel of a selected region, producing a map of the sample’s deformation.

a. Graphene atomic lattice. Each ball represents a carbon atom. b. Twisted graphene bilayers. The three main stacking configurations are shown (AA, AB and domain wall). The moiré unit cell is highlighted. c. Microscope tip in contact with the graphene bilayer d. The strain on the graphene layer bends the chemical bonds between the atoms from in-plane (left) to a mixed in-plane/out-of-plane character (right).

In principle, it was not obvious that a moiré pattern would be detectable with the microscope. When moiré patterns form, it creates a repetitive set of individual units that are called unit cells (highlighted region in Fig. 1b). Each unit cell is formed by regions with different atomic three-dimensional configurations, called sites. In the case of graphene, those sites are called AA and AB which stands for how the atoms from each layer lie on top of each other (see insets in Fig. 1b). The AB regions (also called domains) are separated by domain walls, as highlighted in Fig. 1b. McGilly and colleagues show that the voltage signal detected with the microscope is localized on the domain walls.

When the moiré pattern forms, the atomic layers relax to accommodate it and the layer wrinkles along the domain wall (see right panel in Fig. 1d). Since the microscope is not sensitive to such small deformation, the origin of the detected signal must be electronic. Flat graphene layers have planar bonds, as shown in the left panel of Fig. 1d. However, the curvature of the wrinkle bends the atomic bonds on the graphene layer, which in turn causes an asymmetric distribution of the charge in the vertical direction and gives rise to an out-of-plane polarization (P), which is responsible for the signal measured in the microscope.

The technique designed by McGilly and colleagues has been proven extremely useful for the advancement of the field due to the simplicity of the method and the fact that it allows imaging of any moiré pattern, independently of the nature of the individual units that conform it – that is, whether they are metals, semiconductors or insulators. Being able to image moiré patterns with such an accessible technique will help improve the fabrication process, and having uniform samples is critical since strain gradients can significantly alter the states of matter that emerge in moiré materials.

 

Dr. Leo McGilly is a Postdoctoral Research Fellow in the Physics Department at Columbia University.

How Bouldering keeps urban communities in shape

So, you too enjoy this amazing sport, where people climb over comparatively short distances without any tools, such as ropes or harnesses? Amazing! But, to quote a famous British ensemble, now for something completely different. Today we want to talk about a more serious and urgent topic: flood risks. The recent flood in the New York City area convincingly showed us the risk of flooding in (highly) populated urban areas. Climate change and socioeconomic developments keep on increasing this risk further and further.

Flooding in NYC
Figure 1: Strong rain in New York City transformed parts of it into Venice’s little brother, with less romance but at least 43 death cases.

The United Nations have formulated in their 2030 Agenda for Sustainable Development 17 goals to »[…] stimulate action over the next fifteen years in areas of critical importance for humanity and the planet«. Goal number 11 is »sustainable cities and communities«. But, to properly address a risk it is necessary to adequately analyze and describe it. Current approaches for urban risk analysis mostly lack two important factors: First, they are mainly qualitative but not quantitative. That means, they accurately describe the what of a risk but not how much. We probably all can agree that the information that the biggest crocodile ever found in nature was longer than the biggest giraffe ever was high is much more impressive than the statement, crocodiles can become really big. This demonstrates why quantitative statements are important.

The second problem they do not address properly is the prediction of urban development. They project city growth rather arbitrarily, seldomly incorporating geographical, social or economic factors associated with urbanization. While predictions are difficult, especially about the future, some information exists which can be used as a guide about most probable development cases.

Dr. Mona Hemmati and colleagues tackled both these problems by developing a framework for understanding the interactions between urbanization and flood risks. To do so, they combined four main components: an urban growth module, a hazard module, a risk assessment module and a policy implementation module. The urban growth module is used to achieve a more realistic urban development prediction and the hazard module to generate floodplains. For the risk assessment module the two previous modules are combined while the policy implementation module is used to implement nonstructural strategies, such as for example development zones or taxation variations.

For the framework development the City of Boulder, Colorado, has been chosen as a testbed. Various data such as size, shape, surrounding area or density distribution of the city has been gathered by different sources and used as input parameters for their model.

Their urban growth model has four key features which are used to predict the urbanization process, divided into residential, industrial and commercial & mixed-used occupation. They divide the urban area and surroundings into equally sized cells, the so-called cell space, creating a 2D spatial grid. Each cell can have a cell state, which describes if the cell is developed or not. The neighbourhood of a cell is a factor which can either have an attractive or repulsive effect on the surrounding cells and the transition potential represents the probability of a cell state change for the next time step, defined by different development factors. For the hazard module a development by the Federal Emergency Management Agency was used. With this different floodplains characteristics can be calculated for various flood scenarios, such as for 5, 10, etc. year return period. The risk assessment module measures the damage to physical infrastructure and caused by economic and social disruptions as expected annual damage (EAD) in $US. Last, the policy implementation module takes into account nonstructural flood mitigation measures. Structural measures, such as for example dams, aim at controlling the hazard and keeping the flood out, while nonstructural measures, such as for example land acquisition or public awareness programmes aim at reducing the exposure to hazard.

Using this framework, they tested two different policies against both the current development policy of the city as well as no policy at all. For the first policy they defined low-risk zones and disallowed development and high-risk zones, while for the second they defined socioeconomic incentives, such as for example placing school and places of entertainment in low-risk zones. The interesting result was, that from the four tested cases, Boulder’s current development policy showed the worst result in terms of growth inside the floodplains and therefore long-term costs. Even uncontrolled development was better, while the best policy was the zoning policy, closely followed by the incentive policy.

It can be summarised that while their model still contained many educated guesses and assumptions and for example neglected the influence of the growth module onto the hazard module it can be considered a huge step forward in comparison to purely qualitative models based on random development. The testbed Boulder showed it can be directly applied to community planners in assisting the mitigation of risks due to future hazards, bringing the science out of their ivory tower into the heart of modern society: The city itself.

Dr. Mona Hemmati is a Postdoctoral Research Scientist in the department for Ocean and Climate Physics at the Lamont-Doherty Earth Observatory (LDEO) of Columbia University.

The Different Perceptions of Cultural Appropriation

The term cultural appropriation, is by far a familiar one. It is defined as situations where a person associated with one group uses cultural elements from another group. These elements can include cultural items like “symbols, genres, expressions, technology and artifacts”. While the term is widely used, actual empirical data surrounding the perception of cultural appropriation is limited. In a recent publication, Dr. Ariel Mosley, a Columbia postdoc, and Dr. Biernat venture into the perception of cultural appropriation. To understand how cultural appropriation is perceived by different groups, Dr. Mosley utilizes an approach of a majority and a minority group in the same community and identifies how each group views different actions as cultural appropriation. 

This study uses multiple examples of cultural appropriation (Figure) to identify the perception of appropriation (whether the example is actually cultural appropriation), perception of harm (whether the appropriation can be harmful to the group the cultural aspect was borrowed from), perception of intent (whether the appropriation was done purposefully), and distinctiveness threat (whether the appropriation threatened cultural aspects that allow the minority group to be distinct from the majority group).

To fully identify the perception of cultural appropriation, this study was divided into five sub-studies. Studies one through three focused on the perception of appropriation, harm and intent, study four focused on manipulating distinctiveness threat, and study five focused on fully crossing the actor and race. They recruited an equal number of adults that either identified as Black or White Americans, with White Americans being considered representative of the majority group and Black Americans representative of the minority group. For studies one through three, the authors set out to answer whether Black Americans or White Americans would have higher perceptions of appropriation, harm, intent, and distinctiveness threat. They used a design where the participants would read scenarios, adopted from social media and news clips, of potential cultural appropriation. In these scenarios the perpetrator, the person doing the appropriating, could be either white or black. The participants were asked to review six possible cases of cultural appropriation (Figure). Throughout the three studies they found that Black participants perceived more cases of appropriation than White participants when the perpetrator was White. In a similar pattern, Black participants saw the scenarios as more harmful, and with intent when the perpetrator was White. When the perpetrator was Black neither White participants nor Black participants saw the scenario as appropriation. Black participants also overall felt an increased distinctiveness threat when compared to White participants. These findings supported Dr. Mosley and Dr. Biernats’ original hypothesis of cultural appropriation being more likely to be perceived when perceivers were members of the minority group.

Since in studies one through three, Black participants felt an increased distinctiveness threat, Dr. Mosley and Dr. Beirnat wanted to see whether increased distinctiveness threat in particular could alter the perception of cultural appropriation. To test this the authors primed the participants in a fourth study for increased distinctiveness threat and focused on one scenario category, “hairstyle” (Figure). They primed the participants to either have increased distinctiveness of threat by having them read, “The Disappearing Color Line in America” or normal distinctiveness of threat by having them read, “The Geography and Climate in America”. Black participants were widely unaffected by the priming with the results mimicking studies one through three, but for White participants, those that were primed for distinctiveness of threat saw the White perpetrators’ actions as cultural appropriation. These results indicated that the level of distinctiveness threat experienced increases the perception of cultural appropriation.

Figure: Detailed depiction of the study designs and categories of cultural appropriation.

Then in study five, to reassure their results, the authors paired a perpetrator with a product that was distinctly part of the participant’s culture. The previous four studies used an item/product that was outside of the perpetrator’s culture, but not necessarily an item belonging to the participant’s culture.  Here they used an item/product that was explicitly part of the participant’s culture. The perpetrator was either a White waiter serving culturally Black cuisines or a Black waiter serving culturally White cuisines. Mimicking their previous studies, they found that Black volunteers were more likely to see cultural appropriation when the waiter was White.

Overall, their study indicated that majority and minority groups perceive cultural appropriation differently, with the minority group being more sensitive to actions that can be perceived as appropriative. They also found that harm and intent correlated with appropriation leading them to the conclusion that both perceptions are part of the appropriation construct. These findings supported their initial hypothesis that power relations and social constructs affect the perception of cultural appropriation and added empirical data to a topic often spoken about but yet understudied.

While Dr. Mosley and Dr. Beirnat have added a significant amount of empirical information on how cultural appropriation is perceived, there is still more to explore. Future studies could expand on how cultural appropriation affects multiple other groups including individuals across different races, sexual orientations, genders and individuals with disabilities. 

 

Dr. Ashley Mosley is a Post-Doctoral Research Scientist in the Department of Psychology at Columbia University. Her research focuses on social cognition, social identity and intergroup biases. More information about Dr. Mosley can be found on her website.

What happens when macrophages refuse to eat the dead?

Macrophages, a type of immune cells, are an integral part of our body’s defense system. The term macrophage comes from two Greek words – makro meaning big and phagien meaning eat, which makes them the “big eaters”. And boy, do they love to eat! Some things that they like to chomp on include bacteria and other foreign substances, dying and dead cells, and cancer cells, thus, acting as the body’s cleanup system. This process of eating is not only important for defending against foreign pathogens but is also essential for cleaning up cell debris and maintaining normal bodily functions.

Macrophages typically encapsulate their food by surrounding it with cell extensions, then engulf it and digest it. Check out some cool videos of macrophages eating some bacteria here. This process of eating is typically called “phagocytosis”. However, the term for macrophages eating dying cells is called “efferocytosis”. This term is derived from the Latin word efferre which translates to “take to the grave” or “to bury”. When this mechanism of disposal of cellular corpses goes wrong, the rotting dead cells can lead to inflammation that damages the surrounding tissue. This can lead to many diseases, including coronary artery disease, chronic obstructive pulmonary disease, cystic fibrosis, and rheumatoid arthritis. In a recent publication from the Tabas lab, Dr. Kasikara and Columbia postdoc Dr. Schilperoort explore the molecular mechanisms that underlie impaired efferocytosis and how that leads to the formation of dangerous plaques in the arteries that supply blood to your heart. The buildup of these plaques leads to a condition called coronary artery disease which remains the leading cause of deaths in the United States, causing about 1 in 4 deaths.

Significant advances in genomic sequencing in the past few years have led to the discovery of several mutations that are often correlated with the occurrence of coronary artery disease in patients. One of these mutations is in a gene encoding a protein called PHACTR1. However, because the mutation is present in a part of the gene outside of where the protein-coding sequence lies, it was unclear if this mutation disrupted efferocytosis by disrupting the function of PHACTR1. PHACTR1 regulates the ability of various cell types to expand, contract, and move. While the ability of macrophages to execute these motions is required to engulf or eat cells, whether PHACTR1 is involved in this process in macrophages and thereby macrophage efferocytosis was not known. In this study, the authors made two important discoveries. Firstly, they found that PHACTR1 is essential for macrophage efferocytosis. Secondly, they found that the mutation decreases the expression levels of PHACTR1. The authors investigated more and established that PHACTR1 is important for maintaining an activated version of a motor protein called myosin which is required for cellular movement. Thus, lower levels of PHACTR1 hamper the ability of macrophages to eat dead cells by disrupting cellular movement. This contributes to the buildup of dying cells in our arteries and a consequent increase in the risk of heart attack and stroke.

Fig 1. Model depicting the relationship between efferocytosis and risk of coronary artery disease. Reduced levels of efferocytosis lead to insufficient clearance of dead cells and consequent plaque formation in the arteries. Figure adapted from Kasikara, JCI 2021.

The results from this study provide novel insights into the role of PHACTR1, myosin, and other associated proteins in the pathogenesis and progression of coronary artery disease. Before this study was performed, we only knew that there was a correlation between an increased risk of heart disease and a mutation in PHACTR1 gene. The authors performed rigorous experiments and demonstrated that the mutation changes PHACTR1 production and that this causes heart disease. This information is extremely valuable as it provides a basis for designing future therapies. For example, increasing PHACTR1 production artificially may be an effective strategy for treating coronary artery disease. As defective macrophage efferocytosis is also involved in the pathogenesis of many other diseases, this study has direct implications for the discovery of new treatment paradigms for these diseases as well.

Ancestry connects non-cancer causing mutations in cancer patients

The cause of cancer as a disease has been partly attributed to genetics across a diverse range of populations. However, it is unclear whether cancer patients carry additional genetic mutations, also known as variants, in non-cancer causing genes and if these variants are evolutionarily related. Because ancestry-specific variants were more recently generated in evolutionary time, they could have been easily missed in analyses where all patients were cumulatively analyzed without consideration for ancestry. A recent concept proposed by geneticists suggests that people are more likely to develop or be protected from diseases based on recently acquired mutations and are less so due to more distant mutations. This is an interesting theory that scientists can now test using genome information from more than 10,000 cancer patients whose ancestries are known. So far, how mutations affect gene expression – whether they completely abolish the expression of gene products (e.g. protein) or result in the creation of a misshapen protein, have only been reported for variants present in patients with European ancestry. The remaining ancestries are yet to be explored.

 

Advances in sequencing technology have made it easier for researchers to access genome sequencing information under clinical settings and for healthcare providers to share personalized diagnoses as part of ‘genomic medicine’ to patients. Using publicly available genome sequencing data for cancer patients, Dr. Xiao Fan and colleagues analyzed the variants in non-cancer causing genes and in “medically actionable” genes in 10,389 cancer patients. The authors found 1.46 billion mutations, which were then filtered through rigorous quality testing of sequencing information followed by expert geneticist review, resulting in a final total reliable set of 2,920 non-cancer related pathogenic and likely pathogenic variants. About 750 of these variants were harbored on average within a quarter of the cancer cases, no matter the heritage. A surprising majority (~27%) of the total variants were displayed in patients with European ancestry, followed sequentially by Latinx/Native American (15%), African American (13%) and East Asian (12%) patients.

 

Because genetic mutations can affect expression of proteins, the authors then dug deep into the variant data to examine whether these variants behaved in an expected manner on a molecular level. When genes contain mutations that cause the protein it encodes to be a shorter version of itself, the mutation is referred to cause a protein “truncation”. Sometimes, a truncating mutation in a gene can trigger a decrease in expression at the messenger RNA (mRNA) level even before the mRNA is used to make the protein. To find out if the variants that produced truncated gene products underwent changes at the mRNA level, the authors measured the gene expression levels of such variants. Of the variants that showed a meaningful difference in gene expression compared to non-cancer patients, a large majority of variants showed a decrease in expression. This result indicated to the authors that truncation-causing variants often work at the mRNA level even before the cells spend energy to make the disease-associated proteins. The authors then examined the behavior of gene variants that do not cause truncations but rather cause just a single swap in the gene sequence, known as “missense” variants. Missense mutations typically only cause a change in one or two building blocks of the protein but do not affect the abundance of the protein itself. Surprisingly, the authors found that the missense variants in their data are unusually regulated in the cancer patients at the mRNA level resulting in a decrease in gene, and therefore, protein expression. This is an uncommon observation, making the authors speculate that missense variants are perhaps controlled by gene-expression independent mechanisms within the cancer patients’ cells.

This study provides a testament to the power of genomic medicine that can be used to complement conventional medical treatment. With a strong sample of ~10,000 cancer patients, this report stands as one of the most comprehensive studies that considers race and ancestry in its analysis. While genomic profiling is becoming more common in medical diagnoses, this study further provides a reason for understanding diseases and invention of medicine based on race, ethnicity and genetic heritage.

Identifying a potential risk factor for alcohol abuse among victims of violence in childhood

Half of all children in the United States have been physically assaulted in their lifetime, according to a 2014 study. This finding is alarming, especially considering that childhood maltreatment and abuse can lead to numerous negative mental health outcomes. 

Researchers and medical professionals around the globe often focus on adverse childhood experiences and their detrimental effects on development. Experiencing threat and violence is frequently correlated with a decreased ability to effectively handle negative emotions and heightened emotional reactivity relative to those who have not experienced such trauma. For instance, a typical situation such as having your toy taken away by a peer in school might invoke an explosive, angry response from a child who has been a victim of abuse. Moreover, research demonstrates that children who have faced abuse are also more likely than others to interpret ambiguous actions (such as a classmate accidentally bumping into them in the hallway) as confrontational.

How might those who’ve experienced violence in their childhood, and also have trouble dealing with negative emotions, respond to everyday stressors (i.e. getting through hard homework sets or dealing with long waits for customer service on the phone)? Dr. Charlotte Heleniak and her colleagues studied this response, called distress tolerance, in a newly published paper.

Levels of distress tolerance vary among different individuals. Someone with low distress tolerance is extraordinarily uncomfortable in situations where they’re facing a challenging obstacle, upset, or experiencing negative emotions that can make it hard to persist in the face of difficulty. They have a harder time working through these difficult events compared to people with higher distress tolerance. Research also shows that people with low distress tolerance may find it necessary to escape bad feelings by seeking immediate relief. This relief can often take the form of substance abuse. 

Additionally, while little research has been done, distress tolerance may make an individual more vulnerable to other mental health problems such anxiety and depression. Because of this, Dr. Heleniak’s team examined whether low distress tolerance is associated with these two mental illnesses, as well as alcohol abuse.

teens drinking
Image from Pixabay

Propensity toward problematic alcohol use in adolescents involves many environmental risk factors such as sociodemographic factors and parental drinking behavior. These can be difficult or impossible to address therapeutically. However, if distress tolerance is indeed tied to substance abuse, this may offer a clearer path toward crafting a psychological intervention. 

Dr. Heleniak and her colleagues studied 287 16- to 17-year-old participants across a broad range of socioeconomic backgrounds. They asked the teens about previous violence exposure in their personal life, and assessed depression, anxiety, and alcohol use. Four months later, they reassessed these parameters.

To examine the teens’ distress tolerance, the researchers used a measure called the Paced Auditory Serial Addition Task, which measures a person’s persistence on a difficult task. The sooner a participant decides to terminate the task, the lower their distress tolerance. The team found that those teens who experienced a heighted amount of violence did indeed have lower distress tolerance. At the initial time point, lower levels of distress tolerance were not associated with any of the three psychopathologies (i.e. alcohol abuse, anxiety, depression).

However, the researchers found that low distress tolerance predicted alcohol abuse from the first time point to the second, about 4 months later. Low distress tolerance was not associated with anxiety or depression at either of the two time points of data collection.

Figure 1. Teens who experienced more abuse and violence had lower distress tolerance. Four months after the initial assessment, teens who had low distress tolerance were even more likely to have developed problematic drinking behaviors.

Based on their findings, Dr. Heleniak and her team conclude that researchers could potentially pinpoint distress tolerance as a way to target teens’ problematic use of alcohol, especially those who have experienced violence. Indeed, therapeutic programs aimed at improving distress tolerance already exist. The authors explain that treatments such as Dialectical Behavior Therapy (DBT) and mindfulness practices may be particularly useful. 

Given that teen alcohol abuse may continue into adulthood and lead to dependency issues later in life, the findings of this study could go a long way to helping those adolescents who struggle with both addiction issues and an abusive past.

If you or someone you know is experiencing substance dependency problems, SAMHSA (1-800-662-HELP) is a free, confidential, resource available 24/7 365-days a year.

Charlotte Heleniak is a postdoctoral scientist in the Developmental Affective Neuroscience Lab at Columbia University. She received her Ph.D. in Child Clinical Psychology from the University of Washington. She focuses on how childhood adversity impacts emotion regulation and social cognition in ways that predict adolescent psychopathology. This research has earned her awards from the National Institute of Mental Health and the Doris Duke Charitable Foundation, as well as the Sparks Early Career Grant from the American Psychological Foundation.