Hostname: page-component-6bf8c574d5-9nwgx Total loading time: 0 Render date: 2025-02-23T13:58:39.843Z Has data issue: false hasContentIssue false

Numerical simulation of turbulent drag reduction using micro-bubbles

Published online by Cambridge University Press:  08 October 2002

JIN XU
Affiliation:
Division of Applied Mathematics, Brown University, Providence, RI 02912, USA
MARTIN R. MAXEY
Affiliation:
Division of Applied Mathematics, Brown University, Providence, RI 02912, USA
GEORGE EM KARNIADAKIS
Affiliation:
Division of Applied Mathematics, Brown University, Providence, RI 02912, USA
Rights & Permissions [Opens in a new window]

Abstract

Core share and HTML view are not available for this content. However, as you have access to this content, a full PDF is available via the ‘Save PDF’ action button.

While turbulent drag reduction through the injection of micro-bubbles into a turbulent boundary layer is well established in experiments, there is a lack of corresponding supporting evidence from direct numerical simulations. Here we report on a series of numerical simulations of small bubbles seeded in a turbulent channel flow at average volume fractions of up to 8%. These results show that even for relatively large bubbles, an initial transient drag reduction can occur as bubbles disperse into the flow. Relatively small spherical bubbles will produce a sustained level of drag reduction over time.

Type
Research Article
Copyright
© 2002 Cambridge University Press