url
stringlengths
14
2.42k
text
stringlengths
100
1.02M
date
stringlengths
19
19
metadata
stringlengths
1.06k
1.1k
http://mathonline.wikidot.com/limits-of-functions-of-three-variables
Limits of Functions of Three Variables We have just looked at Limits of Functions of Two Variables. Recall that for a two variable real-valued function, $z = f(x, y)$, then $\lim_{(x, y) \to (a,b)} f(x, y) = L$ if $\forall \epsilon > 0$ $\exists \delta > 0$ such that if $x \in D(f)$ and $0 < \sqrt{(x - a)^2 + (y - b)^2} < \delta$ then $\mid f(x,y) - L \mid < \epsilon$. We are now going to extend this concept further to functions of three variables. Definition: Let $w = f(x, y, z)$ be a three variable real-valued function. Then the Limit of $f(x, y, z)$ as $(x, y, z)$ Approaches $(a,b,c)$ is $L$ denoted $\lim_{(x, y, z) \to (a,b,c)} f(x, y, z) = L$ if $\forall \epsilon > 0$ $\exists \delta > 0$ such that if $(x, y, z) \in D(f)$ and $0 < \sqrt{(x - a)^2 + (y - b)^2 + (z - c)^2} < \delta$ then $\mid f(x, y, z) - L \mid < \epsilon$. Once again, it is important to note that $\sqrt{(x - a)^2 + (y - b)^2 + (z - c)^2}$ represents the distance between the points $(x, y, z)$ and $(a, b, c)$. Thus, we can reformulate the definition above as follows. The limit as $(x, y, z)$ approaches $(a, b, c)$ is the real number $L$ if for all $\epsilon > 0$ there exists a $\delta > 0$ such that if the distance between the points $(x, y, z)$ and $(a, b, c)$ is less than $\delta$ but not $0$, then the distance between $f(x, y, z)$ and $L$ is less than $\epsilon$. We can also look at limits of functions of more than $3$ variables: Definition: Let $z = f(x_1, x_2, ..., x_n)$ be an $n$ variable real-valued function. Then the Limit of $f(x_1, x_2, ..., x_n)$ as $(x_1, x_2, ..., x_n)$ Approaches $(a_1, a_2, ..., a_n)$ is $L$ denoted $\lim_{(x_1, x_2, ..., x_n) \to (a_1, a_2, ..., a_n)} f(x_1, x_2, ..., x_n) = L$ if $\forall \epsilon > 0$ $\exists \delta > 0$ such that if $(x_1, x_2, ..., x_n) \in D(f)$ and $0 < \sqrt{(x_1 - a_1)^2 + (x_2 - a_2)^2 + ... + (x_n - a_n)^2} < \delta$ then $\mid f(x_1, x_2, ..., x_n) - L \mid < \epsilon$.
2020-06-01 20:07:06
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9860774874687195, "perplexity": 29.246770470141737}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-24/segments/1590347419593.76/warc/CC-MAIN-20200601180335-20200601210335-00154.warc.gz"}
http://bsdupdates.com/error-propagation/propagate-error.php
Home > Error Propagation > Propagate Error # Propagate Error ## Contents R., 1997: An Introduction to Error Analysis: The Study of Uncertainties in Physical Measurements. 2nd ed. References Skoog, D., Holler, J., Crouch, S. Now a repeated run of the cart would be expected to give a result between 36.1 and 39.7 cm/s. This is the most general expression for the propagation of error from one set of variables onto another. useful reference In this case, expressions for more complicated functions can be derived by combining simpler functions. The fractional indeterminate error in Q is then 0.028 + 0.0094 = 0.122, or 12.2%. Or in matrix notation, f ≈ f 0 + J x {\displaystyle \mathrm σ 6 \approx \mathrm σ 5 ^ σ 4+\mathrm σ 3 \mathrm σ 2 \,} where J is We say that "errors in the data propagate through the calculations to produce error in the result." 3.2 MAXIMUM ERROR We first consider how data errors propagate through calculations to affect Read More Here ## Error Propagation Calculator Then the displacement is: Dx = x2-x1 = 14.4 m - 9.3 m = 5.1 m and the error in the displacement is: (0.22 + 0.32)1/2 m = 0.36 m Multiplication Measurements Lab 21.845 προβολές 5:48 Uncertainty estimates for physics labs - Διάρκεια: 14:26. Summarizing: Sum and difference rule. Khan Academy 501.848 προβολές 15:15 Error propagation - Διάρκεια: 10:29. For example, the rules for errors in trigonometric functions may be derived by use of the trigonometric identities, using the approximations: sin θ ≈ θ and cos θ ≈ 1, valid are inherently positive. If you measure the length of a pencil, the ratio will be very high. Error Propagation Definition Resistance measurement A practical application is an experiment in which one measures current, I, and voltage, V, on a resistor in order to determine the resistance, R, using Ohm's law, R ISBN0470160551.[pageneeded] ^ Lee, S. In matrix notation, [3] Σ f = J Σ x J ⊤ . {\displaystyle \mathrm {\Sigma } ^{\mathrm {f} }=\mathrm {J} \mathrm {\Sigma } ^{\mathrm {x} }\mathrm {J} ^{\top }.} That When two quantities are divided, the relative determinate error of the quotient is the relative determinate error of the numerator minus the relative determinate error of the denominator. We'd have achieved the elusive "true" value! 3.11 EXERCISES (3.13) Derive an expression for the fractional and absolute error in an average of n measurements of a quantity Q when IIT-JEE Physics Classes 834 προβολές 8:52 Measurements, Uncertainties, and Error Propagation - Διάρκεια: 1:36:37. Error Propagation Inverse General functions And finally, we can express the uncertainty in R for general functions of one or mor eobservables. You can change this preference below. Κλείσιμο Ναι, θέλω να τη κρατήσω Αναίρεση Κλείσιμο Αυτό το βίντεο δεν είναι διαθέσιμο. Ουρά παρακολούθησηςΟυράΟυρά παρακολούθησηςΟυρά Κατάργηση όλωνΑποσύνδεση Φόρτωση... Ουρά παρακολούθησης Ουρά __count__/__total__ Propagation See Ku (1966) for guidance on what constitutes sufficient data. 1. This forces all terms to be positive. 2. The exact covariance of two ratios with a pair of different poles p 1 {\displaystyle p_{1}} and p 2 {\displaystyle p_{2}} is similarly available.[10] The case of the inverse of a 3. Wolfram Demonstrations Project» Explore thousands of free applications across science, mathematics, engineering, technology, business, art, finance, social sciences, and more. 4. Journal of Research of the National Bureau of Standards. ## Error Propagation Physics It's a good idea to derive them first, even before you decide whether the errors are determinate, indeterminate, or both. Define f ( x ) = arctan ⁡ ( x ) , {\displaystyle f(x)=\arctan(x),} where σx is the absolute uncertainty on our measurement of x. Error Propagation Calculator These modified rules are presented here without proof. Error Propagation Chemistry Journal of the American Statistical Association. 55 (292): 708–713. In either case, the maximum size of the relative error will be (ΔA/A + ΔB/B). see here However, in complicated scenarios, they may differ because of: unsuspected covariances disturbances that affect the reported value and not the elementary measurements (usually a result of mis-specification of the model) mistakes Data Reduction and Error Analysis for the Physical Sciences. Retrieved 22 April 2016. ^ a b Goodman, Leo (1960). "On the Exact Variance of Products". Error Propagation Square Root Young, V. It is a calculus derived statistical calculation designed to combine uncertainties from multiple variables, in order to provide an accurate measurement of uncertainty. When mathematical operations are combined, the rules may be successively applied to each operation. this page In other classes, like chemistry, there are particular ways to calculate uncertainties. The fractional determinate error in Q is 0.028 - 0.0094 = 0.0186, which is 1.86%. Error Propagation Excel Note this is equivalent to the matrix expression for the linear case with J = A {\displaystyle \mathrm {J=A} } . It is important to note that this formula is based on the linear characteristics of the gradient of f {\displaystyle f} and therefore it is a good estimation for the standard ## Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc., a non-profit organization. doi:10.1007/s00158-008-0234-7. ^ Hayya, Jack; Armstrong, Donald; Gressis, Nicolas (July 1975). "A Note on the Ratio of Two Normally Distributed Variables". Anytime a calculation requires more than one variable to solve, propagation of error is necessary to properly determine the uncertainty. Therefore the fractional error in the numerator is 1.0/36 = 0.028. Error Propagation Average To contrast this with a propagation of error approach, consider the simple example where we estimate the area of a rectangle from replicate measurements of length and width. However, in complicated scenarios, they may differ because of: unsuspected covariances errors in which reported value of a measurement is altered, rather than the measurements themselves (usually a result of mis-specification External links A detailed discussion of measurements and the propagation of uncertainty explaining the benefits of using error propagation formulas and Monte Carlo simulations instead of simple significance arithmetic Uncertainties and One drawback is that the error estimates made this way are still overconservative. http://bsdupdates.com/error-propagation/propagate-error-mean.php Raising to a power was a special case of multiplication. There is no error in n (counting is one of the few measurements we can do perfectly.) So the fractional error in the quotient is the same size as the fractional Guidance on when this is acceptable practice is given below: If the measurements of $$X$$, $$Z$$ are independent, the associated covariance term is zero. A similar procedure is used for the quotient of two quantities, R = A/B. For highly non-linear functions, there exist five categories of probabilistic approaches for uncertainty propagation;[6] see Uncertainty Quantification#Methodologies for forward uncertainty propagation for details. The number "2" in the equation is not a measured quantity, so it is treated as error-free, or exact. Please see the following rule on how to use constants. Then σ f 2 ≈ b 2 σ a 2 + a 2 σ b 2 + 2 a b σ a b {\displaystyle \sigma _{f}^{2}\approx b^{2}\sigma _{a}^{2}+a^{2}\sigma _{b}^{2}+2ab\,\sigma _{ab}} or The coefficients may also have + or - signs, so the terms themselves may have + or - signs. So, rounding this uncertainty up to 1.8 cm/s, the final answer should be 37.9 + 1.8 cm/s.As expected, adding the uncertainty to the length of the track gave a larger uncertainty It is therefore likely for error terms to offset each other, reducing ΔR/R.
2018-01-20 15:11:42
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8469390273094177, "perplexity": 1631.6933734533704}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-05/segments/1516084889660.55/warc/CC-MAIN-20180120142458-20180120162458-00594.warc.gz"}
https://rationalwiki.org/wiki/Morton%27s_fork
# Morton's fork Cogito ergo sumLogic and rhetoric Key articles General logic Bad logic v - t - e Morton's fork is the term given to a tricky bit of sophistry in which a conclusion is drawn in several different ways that contradict each other. It is named after Cardinal John Morton (Henry VII of England's sidekick). Morton was a tax collector. If someone was lavish in their spending it was proof they were rich and could afford to pay what was asked, if they were less forthcoming, it was evidence that they must have a great deal of savings from living such an austere lifestyle, and hence could likewise afford to pay, although in Morton's defense this tactic was primarily intended for prying money out of the clutches of tax-dodging Monopoly Men who were already known to be rich (some things never change, do they?). Variations of this are often used by creationists, who will use any observation, as well as its inverse, as "proof" of Biblical literalism. This is also a favorite of conspiracy theorists, who believe evidence against the conspiracy is actually evidence for the conspiracy because those perpetrating the conspiracy are obviously covering up the necessary evidence. This fallacy can be neatly summed up by the phrase "Heads, I win; Tails, you lose." Eliezer Yudkowsky discusses this fallacy when formulating what he calls the law of conservation of expected evidence: if data D is considered evidence for hypothesis H, then the opposite data ~D must be considered evidence against hypothesis H.[1] In this sense, evidence-based reasoning is different from deductive reasoning (formal logic-based), where ${\displaystyle P\Rightarrow Q}$ does not imply ${\displaystyle \neg P\Rightarrow \neg Q}$. ## Examples ### Fine Tuning Argument The Fine Tuning Argument attempts to prove the existence of a god by showing that the likelihood of a universe arising in which our life is supported is so small that our existence in a universe that supports us shows something must have created the universe for us. Essentially it states that because we are in a universe which can support our life, but it seems likely that most "possible" universes could not support our life, a god must have ensured this universe supports our life. However, if we existed in any of the multitude of "possible" universes that clearly shouldn't support our life, this would then become evidence that some god had somehow protected us so we formed despite all possibilities, rather than the universe forming to support us despite all the possibilities. So either a god exists because we are in one of the few and rare "possible" universes that can support our life without a god, or a god exists because we are in one of the many universes that can't support our life without a god. ### Witchcraft See the main article on this topic: Witchcraft During many periods of history, including the Salem Witch Trials, the European witch-hunts, and the McCarthy Era, the question has been posed to anyone presumed guilty of a crime: if they did confess, then they would be relieved of their belongings (which would, more often than not, end up with their accuser) and set free to die on their own, whereas if they declared themselves innocent, they would be tortured to death, then relieved of their belongings, as a truly guilty person would never confess. This was later applied in Arthur Miller's play, The Crucible, which while set during the period of the Salem Witch Trials, was allegorical for the then-contemporary McCarthyism. Other examples are listed by Friedrich Spee von Langenfeld, a German priest who opposed witch hunts, in his 1631 work Cautio Criminalis. If a woman led an evil and improper life, it was considered proof that she was a witch; if she led a good and proper life, this was also considered evidence that she was a witch, because witches try to appear virtuous.[1] ### In fiction A well-known example occurs in the comedy film Monty Python's Life of Brian when the protagonist, an ordinary Jewish man in 1st century Jerusalem who accidentally gathers a crowd of religious devotees, tries to convince them that he's not the Messiah. Brian: I'm not the Messiah! Will you please listen? I'm not the Messiah! Do you understand? Honestly! Followers: Only the true Messiah denies his divinity. Brian: What? Well, what sort of chance does that give me? All right, I am the Messiah! Followers: He is! He is the Messiah! ## Alternate Interpretation The term "Morton's Fork" is also sometimes used more broadly to refer to any situation where one party can choose between a number of courses of action, none of which have a favorable result. I.e. "damned if you do, damned if you don't." Examples: • The Nuremburg Defense ("I was just following orders") has sometimes been, well, defended by pointing out that the soldiers using it were put in such a situation the instant they were given those orders: Eithr they could maybe be punished later for committing a war crime if they lost, or they could definitely be punished now for, you know, disobeying a direct order. • In 2011 a twenty year old woman claimed she'd been knocked up by Bieber the year before and sued him for compensation. The dumbass would've been sent to jail regardless of how the case turned out, since Bieber would've been underage at the time: if she lost, for fraud; if she won, for having sex with a minor. The case quietly went away. • Woomeisters who flaunt "the world will end on this specific date" predictions tend to realize all too late that they've impaled themselves on Morton's long, pointy fork. Either they turn out to be right, in which case they're probably dead along with everyone else, or they turn out to be wrong, in which case their credibility is permanently ruined.
2022-10-04 13:38:00
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 2, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.384491503238678, "perplexity": 2787.760422622671}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337504.21/warc/CC-MAIN-20221004121345-20221004151345-00193.warc.gz"}
http://descartes.math.psu.edu/calendars/meeting.php?id=31735
# Meeting Details Title: Derived stacks for dummies: Mk. II Student Geometric Functional Analysis Seminar Damien Broka, Penn State http:// During the first talk, we managed to convince a very legitimately worried audience to consider an extremely over-the-top definition of sheaves on manifolds. We promised this would prove to be the right language to launch ourselves into the descent condition for $\infty$-stacks. In this session, "Mk. II", we're going for the (derived $\infty$-)moon baby. Or we'll die trying.
2018-01-19 14:06:13
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4513744115829468, "perplexity": 8064.027592578511}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-05/segments/1516084887981.42/warc/CC-MAIN-20180119125144-20180119145144-00617.warc.gz"}
http://www.numdam.org/item/?id=CM_1993__89_2_241_0
Corrections to “How to explicitly solve a Thue-Mahler equation” Compositio Mathematica, Tome 89 (1993) no. 2, pp. 241-242. @article{CM_1993__89_2_241_0, author = {Tzanakis, N. and de Weger, Benjamin M. M.}, title = {Corrections to {\textquotedblleft}How to explicitly solve a Thue-Mahler equation{\textquotedblright}}, journal = {Compositio Mathematica}, pages = {241--242}, } Tzanakis, N.; de Weger, B. M. M. Corrections to “How to explicitly solve a Thue-Mahler equation”. Compositio Mathematica, Tome 89 (1993) no. 2, pp. 241-242. http://www.numdam.org/item/CM_1993__89_2_241_0/
2021-10-23 08:45:46
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.27825242280960083, "perplexity": 7258.706961621472}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-43/segments/1634323585653.49/warc/CC-MAIN-20211023064718-20211023094718-00452.warc.gz"}
http://www.emathzone.com/tutorials/math-results-and-formulas/formulas-and-results-of-permutations-combinations-and-binomial-theorem.html
# Formulas and Results of Permutations Combinations and Binomial Theorem 1. Factorial of $n$ is defined as $n! = n \cdot (n - 1) \cdot (n - 2) \cdot (n - 3) \ldots 3 \cdot 2 \cdot 1$, where is natural number. 2. ${}^n{P_r} = \frac{{n!}}{{(n - r)!}}$ (Permutation) 3. ${}^n{C_r} = \left( {\begin{array}{*{20}{c}} n \\ r \end{array}} \right) = \frac{{n!}}{{r!(n - r)!}}$ (Combination) 4. $\left( {\begin{array}{*{20}{c}} n \\ r \end{array}} \right) = \left( {\begin{array}{*{20}{c}}n \\ {n - r} \end{array}} \right)$ 5. $\left( {\begin{array}{*{20}{c}}n \\ r \end{array}} \right) = \left( {\begin{array}{*{20}{c}}n \\ {r + 1} \end{array}} \right) = \left( {\begin{array}{*{20}{c}} {n + 1} \\ {r + 1} \end{array}} \right)$ $\forall n,r \in \mathbb{N}$ 6. ${}^n{C_r} = \frac{{{}^n{P_r}}}{{r!}}$ 7. $\left( {\begin{array}{*{20}{c}}2 \\ 2 \end{array}} \right) + \left( {\begin{array}{*{20}{c}}3 \\ 2 \end{array}} \right) + \left( {\begin{array}{*{20}{c}} 4 \\ 2 \end{array}} \right) + \cdots + \left( {\begin{array}{*{20}{c}} {n + 1} \\ 2 \end{array}} \right) = \left( {\begin{array}{*{20}{c}} {n + 2} \\ 3 \end{array}} \right)$ 8. For large $n$, $n! = \sqrt {2\pi n} {\text{ }}{n^n}{\text{ }}{e^{ - n}}$ is called Striling Approximation Formula. 9. $\frac{n}{2} = \left( {\begin{array}{*{20}{c}}n \\ 0 \end{array}} \right) + \left( {\begin{array}{*{20}{c}}n \\ 1 \end{array}} \right) + \left( {\begin{array}{*{20}{c}} n \\ 2 \end{array}} \right) + \cdots + \left( {\begin{array}{*{20}{c}}n \\ n \end{array}} \right)$ 10. $\left( {\begin{array}{*{20}{c}}n \\ 0 \end{array}} \right) = 1$, $\left( {\begin{array}{*{20}{c}}n \\ 1 \end{array}} \right) = n$, $\left( {\begin{array}{*{20}{c}} n \\ 2 \end{array}} \right) = \frac{{n(n - 1)}}{{2!}}$, $\left( {\begin{array}{*{20}{c}}n \\ 3 \end{array}} \right) = \frac{{n(n - 1)(n - 2)}}{{3!}}$ 11. $\left( {\begin{array}{*{20}{c}}n \\ {n - 1} \end{array}} \right) = n$, $\left( {\begin{array}{*{20}{c}} n \\ n \end{array}} \right) = 1$ 12. If $n$ is positive integer, then ${(a + b)^n} = \left( {\begin{array}{*{20}{c}}n \\ 0 \end{array}} \right){a^n} + \left( {\begin{array}{*{20}{c}} n \\ 1 \end{array}} \right){a^{n - 1}}b + \left( {\begin{array}{*{20}{c}} n \\ 2 \end{array}} \right){a^{n - 2}}{b^2} + \cdots + \left( {\begin{array}{*{20}{c}}n \\ r \end{array}} \right){a^{n - r}}{b^r} + \cdots + \left( {\begin{array}{*{20}{c}}n \\ n \end{array}} \right){b^n}$ 13. $\begin{gathered}{(a + b)^n} = {a^n} + n{a^{n - 1}}b + \frac{{n(n - 1)}}{{2!}}{a^{n - 2}}{b^2} + \cdots \\ \cdots + \frac{{n(n - 1)(n - 2) \cdots (n - r + 1)}}{{r!}}{a^{n - r}}{b^r} + \cdots + {b^n} \\ \end{gathered}$ 14. General Term$= \left( {\begin{array}{*{20}{c}}n \\ r \end{array}} \right){a^{n - r}}{b^r}$ 15. If $n$ is not positive integer, then the binomial expansion is ${(1 + x)^n} = 1 + nx + \frac{{n(n - 1)}}{{2!}}{x^2} + \frac{{n(n - 1)(n - 2)}}{{3!}} + \cdots$ is called the binomial series.
2017-01-24 19:18:08
{"extraction_info": {"found_math": true, "script_math_tex": 24, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 24, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9673535823822021, "perplexity": 990.8841708073178}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-04/segments/1484560285244.23/warc/CC-MAIN-20170116095125-00358-ip-10-171-10-70.ec2.internal.warc.gz"}
https://homotopytypetheory.org/2021/01/23/can-the-type-of-truth-values-be-given-the-structure-of-a-group/
## Can the type of truth values be given the structure of a group? When attempting to prove a theorem, not discussed in this post, I tried to give a certain set the structure of a group. After failing repeatedly, my proof attempts led me to consider a miniature version of the problem: can $\Omega$, the object of truth values, be given the structure of a group? This question makes sense in 1-topos theory, in $\infty$-topos theory and in HoTT/UF, where in the latter a truth value is a type with at most one element, also known as a proposition (and sometimes as an h-proposition or as a mere proposition). The answer to the question, in all cases, turns out to be that $\Omega$ can be given the structure of a group if and only if the topos is boolean. Reasoning in the internal language, or the type theory, the argument is as follows. If the topos is boolean, that is, excluded middle holds in the internal language, then trivially $\Omega$ has precisely two group structures, and $\Omega$ endowed with each of them gives rise to two isomorphic groups. It is the converse that is interesting. We use “Higgs’ Involution Theorem”, which says that every left-cancellable map $f : \Omega \to \Omega$ is an involution, that is, $f (f p) = p$. If a group structure $(\oplus , 0)$ on $\Omega$ is given, by Higgs’ involution theorem we conclude that it is abelian with $p \oplus p=0$ for every $p:\Omega$ (where it is unspecified which element of $\Omega$ the zero element of the group is – it can be $\text{false}$, $\text{true}$, or any element of $\Omega$). Now define $f(p) = p \oplus \text{false} \oplus \text{true}$. Then we have the following chain of logical equivalences: $f(p)$ $\iff f(p) = \text{true}$ $\iff p \oplus \text{false} \oplus \text{true} = \text{true}$ $\iff p \oplus \text{false} = 0$ $\iff p = \text{false}$ $\iff \lnot p$. Hence $f(p) = \lnot p$ by propositional extensionality (or univalence for propositions, in the case of HoTT/UF). But $f(p)$, being a group action, is an automorphism and hence an involution by Higgs’ theorem (or already by the above consequences of Higgs’ theorem and direct calculation), which amounts to $\lnot\lnot p = p$, and therefore excluded middle holds. This argument is also written here in Agda, with a weaker assumption, namely that $\Omega$ is given the structure of a left-cancellative monoid, including a proof of Higgs Involution Theorem. I wouldn’t be surprised if this fact is already available in the literature or the folklore. Is it? https://www.cs.bham.ac.uk/~mhe/ This entry was posted in Uncategorized. Bookmark the permalink. ### 7 Responses to Can the type of truth values be given the structure of a group? 1. Update: none of the topos theorists we asked, including Peter Johnstone, were aware of the fact that the object of truth values can be a group only in a boolean topos. Nevertheless, the paper “[Automorphisms of Ω](https://link.springer.com/article/10.1007/BF02488012)” by Johnstone (algebra universalis volume 9, pages1–7 (1979)) is relevant. There he proves the externalization of the fact that the image of the Higgs filter H := { p:Ω | [p,true] = {p,true} } under the map sending p:Ω to p _ is precisely Aut(Ω) = Mono(Ω,Ω), and I believe his arguments actually prove the full-fledged internal version too. Then, a group structure on Ω provides an embedding of Ω into H; composite of this embedding with the inclusion of H into Ω is an embedding of Ω into itself, hence an automorphism, hence the inclusion of H into Ω is epi, hence H contains false, so that Ω = {false, true}. Your argument is doubtlessly more elementary, though. Hm, biimplication vanished by some reason… Actually already a mono $i:\Omega\to\Omega$ with $i(\mathbf{false})=\mathbf{true}$ implies booleanness. Indeed since $i$ is mono, $i(p)=\mathbf{true}$ implies $p=\mathbf{false}$, so that $i(p)=(p=\mathbf{false})$. Then again since $i$ is mono, $i(p)=i(q)$ implies $p=q$. Taking here $q=((p=\mathbf{false})=\mathbf{false})$ gives (since triple negation coincides with single negation) $((p=\mathbf{false})=\mathbf{false})=p$, i. e. $\neg\neg p = p$, which gives booleanness. Thanks for the link! It is remarkable – in a sense our assumptions are opposite to each other: we both deal with a mono $i:\Omega\rightarrowtail\Omega$; you assume $i(1)=0$ and I $i(0)=1$. Conclusions are curiously similar yet different: for me, it is that $i=\neg$, so $latex\neg$ is mono, so one can cancel $\neg\neg\neg=\neg\text{identity}$ from the right, implying that $\neg\neg$ is identity. For you, that $latex\neg i$ is identity, so that $latex\neg$ is (a split) epi, hence booleanness by your Lemma 15. The latter, in effect, cancels $\neg\neg\neg=\text{identity}\neg$ from the left!
2022-05-17 21:56:14
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 33, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9466351866722107, "perplexity": 565.6737469288371}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652662520817.27/warc/CC-MAIN-20220517194243-20220517224243-00314.warc.gz"}
https://www.aimsciences.org/article/doi/10.3934/amc.2016027
# American Institute of Mathematical Sciences August  2016, 10(3): 583-588. doi: 10.3934/amc.2016027 ## There is no $[24,12,9]$ doubly-even self-dual code over $\mathbb F_4$ 1 Department of EE-Systems, Tel Aviv University, Tel Aviv, Israel 2 Lehrstuhl D für Mathematik, RWTH Aachen University, 52056 Aachen Received  May 2015 Revised  October 2015 Published  August 2016 We show that there is no $[24,12,9]$ doubly-even self-dual code over $\mathbb{F}_4$ by attempting to construct the generator matrix of this code directly. Citation: Sihuang Hu, Gabriele Nebe. There is no $[24,12,9]$ doubly-even self-dual code over $\mathbb F_4$. Advances in Mathematics of Communications, 2016, 10 (3) : 583-588. doi: 10.3934/amc.2016027 ##### References: [1] E. F. Assmus, Jr. and H. F. Mattson, Jr., New $5$-designs,, J. Combin. Theory, 6 (1969), 122. Google Scholar [2] K. Betsumiya, On the classification of type II codes over $\mathbbF_{2^r}$ with binary length 32,, preprint., (). Google Scholar [3] K. Betsumiya, T. A. Gulliver, M. Harada and A. Munemasa, On type II codes over $\mathbbF_4$,, IEEE Trans. Inform. Theory, 47 (2001), 2242. doi: 10.1109/18.945245. Google Scholar [4] W. Bosma, J. Cannon and C. Playoust, The Magma algebra system. I. The user language,, J. Symbolic Comput., 24 (1997), 235. doi: 10.1006/jsco.1996.0125. Google Scholar [5] T. Feulner, The automorphism groups of linear codes and canonical representatives of their semilinear isometry classes,, Adv. Math. Commun., 3 (2009), 363. doi: 10.3934/amc.2009.3.363. Google Scholar [6] P. Gaborit, V. Pless, P. Solé and O. Atkin, Type II codes over $\mathbbF_4$,, Finite Fields Appl., 8 (2002), 171. doi: 10.1006/ffta.2001.0333. Google Scholar [7] A. Günther, Codes und Invariantentheorie (in German),, Diploma thesis, (2006). Google Scholar [8] W. C. Huffman and V. Pless, Fundamentals of Error-Correcting Codes,, Cambridge University Press, (2003). doi: 10.1017/CBO9780511807077. Google Scholar [9] G. Nebe, H.-G. Quebbemann, E. M. Rains and N. J. A. Sloane, Complete weight enumerators of generalized doubly-even self-dual codes,, Finite Fields Appl., 10 (2004), 540. doi: 10.1016/j.ffa.2003.12.001. Google Scholar [10] H.-G. Quebbemann, On even codes,, Discrete Math., 98 (1991), 29. doi: 10.1016/0012-365X(91)90410-4. Google Scholar [11] The Sage Development Team, Sage Mathematics Software, Version 6.4.1,, , (). Google Scholar show all references ##### References: [1] E. F. Assmus, Jr. and H. F. Mattson, Jr., New $5$-designs,, J. Combin. Theory, 6 (1969), 122. Google Scholar [2] K. Betsumiya, On the classification of type II codes over $\mathbbF_{2^r}$ with binary length 32,, preprint., (). Google Scholar [3] K. Betsumiya, T. A. Gulliver, M. Harada and A. Munemasa, On type II codes over $\mathbbF_4$,, IEEE Trans. Inform. Theory, 47 (2001), 2242. doi: 10.1109/18.945245. Google Scholar [4] W. Bosma, J. Cannon and C. Playoust, The Magma algebra system. I. The user language,, J. Symbolic Comput., 24 (1997), 235. doi: 10.1006/jsco.1996.0125. Google Scholar [5] T. Feulner, The automorphism groups of linear codes and canonical representatives of their semilinear isometry classes,, Adv. Math. Commun., 3 (2009), 363. doi: 10.3934/amc.2009.3.363. Google Scholar [6] P. Gaborit, V. Pless, P. Solé and O. Atkin, Type II codes over $\mathbbF_4$,, Finite Fields Appl., 8 (2002), 171. doi: 10.1006/ffta.2001.0333. Google Scholar [7] A. Günther, Codes und Invariantentheorie (in German),, Diploma thesis, (2006). Google Scholar [8] W. C. Huffman and V. Pless, Fundamentals of Error-Correcting Codes,, Cambridge University Press, (2003). doi: 10.1017/CBO9780511807077. Google Scholar [9] G. Nebe, H.-G. Quebbemann, E. M. Rains and N. J. A. Sloane, Complete weight enumerators of generalized doubly-even self-dual codes,, Finite Fields Appl., 10 (2004), 540. doi: 10.1016/j.ffa.2003.12.001. Google Scholar [10] H.-G. Quebbemann, On even codes,, Discrete Math., 98 (1991), 29. doi: 10.1016/0012-365X(91)90410-4. Google Scholar [11] The Sage Development Team, Sage Mathematics Software, Version 6.4.1,, , (). Google Scholar [1] Masaaki Harada, Akihiro Munemasa. On the covering radii of extremal doubly even self-dual codes. Advances in Mathematics of Communications, 2007, 1 (2) : 251-256. doi: 10.3934/amc.2007.1.251 [2] Masaaki Harada. New doubly even self-dual codes having minimum weight 20. Advances in Mathematics of Communications, 2020, 14 (1) : 89-96. doi: 10.3934/amc.2020007 [3] Stefka Bouyuklieva, Iliya Bouyukliev. Classification of the extremal formally self-dual even codes of length 30. Advances in Mathematics of Communications, 2010, 4 (3) : 433-439. doi: 10.3934/amc.2010.4.433 [4] Masaaki Harada, Takuji Nishimura. An extremal singly even self-dual code of length 88. Advances in Mathematics of Communications, 2007, 1 (2) : 261-267. doi: 10.3934/amc.2007.1.261 [5] Masaaki Harada, Katsushi Waki. New extremal formally self-dual even codes of length 30. Advances in Mathematics of Communications, 2009, 3 (4) : 311-316. doi: 10.3934/amc.2009.3.311 [6] Gabriele Nebe, Wolfgang Willems. On self-dual MRD codes. Advances in Mathematics of Communications, 2016, 10 (3) : 633-642. doi: 10.3934/amc.2016031 [7] Masaaki Harada, Akihiro Munemasa. Classification of self-dual codes of length 36. Advances in Mathematics of Communications, 2012, 6 (2) : 229-235. doi: 10.3934/amc.2012.6.229 [8] Stefka Bouyuklieva, Anton Malevich, Wolfgang Willems. On the performance of binary extremal self-dual codes. Advances in Mathematics of Communications, 2011, 5 (2) : 267-274. doi: 10.3934/amc.2011.5.267 [9] Nikolay Yankov, Damyan Anev, Müberra Gürel. Self-dual codes with an automorphism of order 13. Advances in Mathematics of Communications, 2017, 11 (3) : 635-645. doi: 10.3934/amc.2017047 [10] Steven T. Dougherty, Joe Gildea, Abidin Kaya, Bahattin Yildiz. New self-dual and formally self-dual codes from group ring constructions. Advances in Mathematics of Communications, 2020, 14 (1) : 11-22. doi: 10.3934/amc.2020002 [11] Hyun Jin Kim, Heisook Lee, June Bok Lee, Yoonjin Lee. Construction of self-dual codes with an automorphism of order $p$. Advances in Mathematics of Communications, 2011, 5 (1) : 23-36. doi: 10.3934/amc.2011.5.23 [12] Bram van Asch, Frans Martens. Lee weight enumerators of self-dual codes and theta functions. Advances in Mathematics of Communications, 2008, 2 (4) : 393-402. doi: 10.3934/amc.2008.2.393 [13] Bram van Asch, Frans Martens. A note on the minimum Lee distance of certain self-dual modular codes. Advances in Mathematics of Communications, 2012, 6 (1) : 65-68. doi: 10.3934/amc.2012.6.65 [14] Katherine Morrison. An enumeration of the equivalence classes of self-dual matrix codes. Advances in Mathematics of Communications, 2015, 9 (4) : 415-436. doi: 10.3934/amc.2015.9.415 [15] Jongmin Han, Juhee Sohn. On the self-dual Einstein-Maxwell-Higgs equation on compact surfaces. Discrete & Continuous Dynamical Systems - A, 2019, 39 (2) : 819-839. doi: 10.3934/dcds.2019034 [16] Minjia Shi, Daitao Huang, Lin Sok, Patrick Solé. Double circulant self-dual and LCD codes over Galois rings. Advances in Mathematics of Communications, 2019, 13 (1) : 171-183. doi: 10.3934/amc.2019011 [17] Steven T. Dougherty, Cristina Fernández-Córdoba, Roger Ten-Valls, Bahattin Yildiz. Quaternary group ring codes: Ranks, kernels and self-dual codes. Advances in Mathematics of Communications, 2019, 0 (0) : 0-0. doi: 10.3934/amc.2020023 [18] Suat Karadeniz, Bahattin Yildiz. New extremal binary self-dual codes of length $68$ from $R_2$-lifts of binary self-dual codes. Advances in Mathematics of Communications, 2013, 7 (2) : 219-229. doi: 10.3934/amc.2013.7.219 [19] Ayça Çeşmelioǧlu, Wilfried Meidl, Alexander Pott. On the dual of (non)-weakly regular bent functions and self-dual bent functions. Advances in Mathematics of Communications, 2013, 7 (4) : 425-440. doi: 10.3934/amc.2013.7.425 [20] Amita Sahni, Poonam Trama Sehgal. Enumeration of self-dual and self-orthogonal negacyclic codes over finite fields. Advances in Mathematics of Communications, 2015, 9 (4) : 437-447. doi: 10.3934/amc.2015.9.437 2018 Impact Factor: 0.879
2019-10-20 01:50:52
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.780494213104248, "perplexity": 7669.238877170123}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570986700560.62/warc/CC-MAIN-20191020001515-20191020025015-00540.warc.gz"}
http://docs.eyesopen.com/oedocking/apopdb2receptor.html
# APOPDB2RECEPTOR¶ ## Overview¶ Converts an apo protein structure into a receptor. The active site is identified by a ligand residue specified by the user (-site_residue). The residue is used as a hint to select most appropriate active site from a list of potential active sites generated by a shape algorithm that searches for pockets/grooves and cavities. Warning The input protein structure must be an apo structure. If the protein structure is in a complex with a ligand use the PDB2RECEPTOR program if the ligand and protein structure are part of a single molecule file, and RECEPTOR_SETUP if they have been separated into two files. ## Input Preparation¶ No special preparation of the protein is required, however the user must identify the name of a residue in active site. ## Parameter List¶ Input -pdb : File with an apo protein structure. -site_residue : Name of a residue in or near the active site of the protein. Receptor Options -solvent_residues : Residue names identifying solvent molecules. Output -receptor : Output receptor filename. Note Values in parenthesis are defaults. ## Parameter Details¶ ### Input¶ -pdb <filename> [No Default: Required Parameter] A file of an apo protein to be converted into a receptor file. Typically this will be a PDB file, however any molecular format with residue information can be used (e.g., OEBinary). Warning The active site must not contain a bound ligand. Use PDB2RECEPTOR to convert a protein-ligand complex into a receptor file. [ Aliases = -in ] -site_residue <residue identifier> [No Default: Required Parameter] The name of a residue near the active site of the protein. Residue names can be of the following forms. Form Example <Residue Name> HIS <Residue Name><Residue Number> HIS1034 <Residue Name><Residue Number><Chain ID> HIS1034A If the residue naming is not specific enough and matches multiple residues in the pdb file the program will list the available, more specific, options to choose from. [ Aliases = -site ] ### Receptor Options¶ -solvent_residues <name> [<name> ...] [Default: HOH] A list of one or more residue names associated with solvent molecules in the pdb file. Molecules that have a residue matching one of these names will be stripped from the protein structure and stored in the receptor as extra molecules (i.e., they will have no affect on docking and scoring). Note The default setting of this flag (HOH) strips waters from the protein. To not strip any molecules specify this flag as ‘NoSolvents’. [ Aliases = -solvent ] ### Output¶ -receptor <filename> [Default: receptor.oeb.gz] Name of the output receptor file to generate. The file must be OEBinary (i.e., either a .oeb or .oeb.gz file). [ Aliases = -rec , -out ] ## Example Commands¶ ### Convert 3E1X into receptor¶ Converts PDB file 3E1X containing an apo protein into a receptor using residue TRP215B to identify the active site. Input files • 3E1X.pdb.gz : PDB file to convert into a receptor. Command line prompt> apopdb2receptor -pdb 3E1X.pdb.gz -site_residue TRP215B Output files • receptor.oeb.gz : Receptor version of 3E1X.
2017-05-23 22:39:37
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4309217929840088, "perplexity": 11021.819577728576}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463607704.68/warc/CC-MAIN-20170523221821-20170524001821-00038.warc.gz"}
https://tex.stackexchange.com/tags/cyrillic/hot
# Tag Info 80 The error you get is due to a "no-break space" character, according to what I can gather by copying an pasting your message. This character is not usually set up by the [utf8] option and it's invisible to many editors, so it can slip in a document without the typist knowing it. Solution: add in your preamble \DeclareUnicodeCharacter{00A0}{ } if you don't ... 60 As this is one of the top google hits for this error message, here's a more general answer with an example: The cause is a unicode character in one of your input files that isn't mapped to an output. This may -- especially if you're using the (unicode-supporting) biblatex/biber system -- be in your bibliography. This is a good place to look for errors as .... 40 For multilingual texts with pdflatex (as opposed to using the xelatex or lualatex engines), the best option is to use the UTF-8 encoding for the input and babel: \documentclass{article} \usepackage[T2A,T1]{fontenc} \usepackage[utf8]{inputenc} \usepackage[russian,english]{babel} \begin{document} Text in English \begin{otherlanguage*}{russian} Текст на ... 20 LaTeX Traditional fonts for LaTeX can have only 256 glyphs, which is clearly insufficient to cover multiple alphabets. This is a problem similar to legacy code pages for operating systems. However, LaTeX can use multiple fonts. The problem has been (not completely) solved with introducing font encodings. Leaving out old and outdated seven bit encodings, ... 20 for someone who wants to use just one or two cyrillic letters in math, in a computer modern setting, the old wncy fonts are still distributed in tex live as part of the amsfonts collection, in type 1 format. (but they don't require use of the amsfonts package.) this is the recommendation in the ams author faq. I want to use some cyrillic letters for ... 19 TeX fonts have only 256 slots. And you can't mix encodings without telling LaTeX to do it; \textcyrillic defined by babel with the russian option does this. The last specified encoding becomes the default one. This should be a complete list of sans serif fonts available also in T2A encoding (for cyrillic): \documentclass{article} \usepackage[T2A,T1]{... 19 Some OpenType fonts support also old Cyrillic letters; compile the following with XeLaTeX or LuaLaTeX: \documentclass{article} \usepackage{fontspec} \setmainfont{Old Standard} \begin{document} \raggedright абвгдежѕзиіклмнопрстѹф% хѡцчшщъꙑьѣюꙗѥѧѫѩѭѯѱѳѵ \uppercase{абвгдежѕзиіклмнопрстѹф хѡцчшщъꙑьѣюꙗѥѧѫѩѭѯѱѳѵ} \end{document} I know no way to get the most ... 18 \usepackage[utf8x]{inputenc} Ubuntu: You must install texlive-latex-extra before use it. Fedora: You must install texlive-collection-latexextra before use it. 18 Doing locate /t2a | grep 'texlive/2012/texmf-dist/tex/latex/.*\.fd\$' outputs, on my updated TeX Live /usr/local/texlive/2012/texmf-dist/tex/latex/antt/t2aantt.fd /usr/local/texlive/2012/texmf-dist/tex/latex/antt/t2aanttc.fd /usr/local/texlive/2012/texmf-dist/tex/latex/antt/t2aanttl.fd /usr/local/texlive/2012/texmf-dist/tex/latex/antt/t2aanttlc.fd /usr/... 16 Edit 10/2017 Here is a new version of the Lua script. It generates the code for Luaotfload, instead of the font feature file, which is not supported anymore. It has also new option --back, which creates mapping in the opposite direction, like from Cyrillic to Latin in our example. The script is named maptolua.lua: kpse.set_program_name "luatex" local lapp ... 15 These characters are available in font encoding T2A, for example. They can be used without babel: \documentclass{article} \usepackage[T2A,T1]{fontenc} \begin{document} U+04B2: {\fontencoding{T2A}\selectfont\CYRHDSC} U+04B3: {\fontencoding{T2A}\selectfont\cyrhdsc} \end{document} In order to avoid PK-Fonts the package cm-super should be installed. 14 Some fonts don't advertise their features correctly, as far as Polyglossia can understand them. In your case you have to announce that Courier New can be used also for cyrillic monospaced bits: \newfontfamily{\cyrillicfonttt}{Courier New} Notice also that if you should enclose English passages in suitable environment, or hyphenation won't be even attempted.... 14 You need to announce LaTeX the languages you intend to use (T2A is for cyrillic): \documentclass{article} \usepackage[T2A,T1]{fontenc} \usepackage[utf8]{inputenc} \usepackage[russian,french,english]{babel} \begin{document} In Dahl’s dictionary there is a similar sounding word “\foreignlanguage{russian}{дуван}”. \begin{otherlanguage*}{french} “C’est ... 12 You can use it directly. fontenc works well with XeTeX/LuaTeX. For example, \documentclass{article} \usepackage[T2A]{fontenc} \begin{document} \symbol{"C1} % produce Б \end{document} However, you can't use it with fontspec. A error will be produced: ! Corrupted NFSS tables. Moreover, it is a problem to input Cyrillic text. inputenc cannot be used in ... 12 The method is similar to that one used for Serbian. Prepare the following cyrillic-to-latin.map file: ; TECkit mapping for TeX input conventions <-> Unicode characters LHSName "Cyrillic-to-Latin" RHSName "UNICODE" pass(Unicode) ; ligatures from Knuth's original CMR fonts U+002D U+002D <> U+2013 ; -- -> en dash U+002D U+002D U+... 12 This is not a problem of cyrillic math characters; if the text were English, the correct input would be Letters only $\text{Memory: } M_{\textup{add}}(n) = \Theta(N)$ because Память and доп are not math. The difference becomes clear when comparing this with the output of $Memory: M_{add}(n) = \Theta(N)$ The bottom formula is clearly wrong. ... 12 this was typeset with luatex: \documentclass{article} \usepackage{fontspec} \usepackage{ifluatex,ifxetex} \setmainfont[Mapping=cyrillic-to-latin]{texgyreadventor-regular.otf} \begin{document} \ifxetex This is XeTeX \else\ifluatex This is LuaTeX \directlua{require("cyrtr")} \fi Hello, world Здравствуй, Мир \end{document} using a cyrtr.lua ... 11 Here is a method for XeLaTeX. Prepare a file ascii-to-serbian.map with the following content: ; TECkit mapping for TeX input conventions <-> Unicode characters LHSName "ASCII-to-Serbian" RHSName "UNICODE" pass(Unicode) ; ligatures from Knuth's original CMR fonts U+002D U+002D <> U+2013 ; -- -> en dash U+002D U+002D U+002D &... 11 Add the line \usepackage[T2A,T1]{fontenc} in the preamble so to have the right font encodings. Thus the MWE \documentclass{letter} \usepackage[utf8]{inputenc} \usepackage[T2A,T1]{fontenc} \usepackage[russian]{babel} \begin{document} \MakeUppercase{English Text. Русский текст.} \end{document} yields the desired result P.S.: The T1 option is not needed ... 11 The character maps of T2A and T1 are identical in the first 128 positions. So english texts should work fine with T2A. But if you want to insert umlauts and other accented chars from the upper part of the char table of T1-encoding you should consider to switch to T1 for such texts instead of relying on LaTeX to fall back to the correct definitions. 10 If you have installed the hyphen-russian and the ruhyphen TeX Live packages, then it can be a problem of non updated formats. From a Terminal window run sudo fmtutil-sys --all and supply your password (assuming you're an administrator of your own machine) when requested. This will recreate all format files and should fix the issue. A note: your example ... 10 There has been some discussion about Cyrillic in math but the issue is still open: https://github.com/wspr/unicode-math/issues/29 As Joseph mentioned you need fonts with the glyph. In case of math (if you don't switch to a text font with \text{..}) you also need to set mathcodes. E.g.; \documentclass[12pt]{book} \usepackage{unicode-math} \setmainfont{Arial ... 9 Your original problem is solved by avoiding Cyrillic in PDF bookmark names. The reason for this is that with inputenc UTF-8 characters are something more than just characters. Instead of \pdfbookmark[0]{Бытие}{Бытие} write \pdfbookmark[0]{Бытие}{Genesis} However, it is easier to use the higher level \addcontentsline command, which not only adds an entry ... 9 Adding to Heiko's excellent answer, one can also input the character directly, provided the file is UTF8 encoded: \documentclass{article} \usepackage[T2A,T1]{fontenc} \usepackage[utf8]{inputenc} \usepackage{newunicodechar} \newunicodechar{Ҳ}{{\fontencoding{T2A}\selectfont\CYRHDSC}} \newunicodechar{ҳ}{{\fontencoding{T2A}\selectfont\cyrhdsc}} \begin{document}... 8 Hmm seems like CJK package is making some definitions global when perhaps it shouldn't. You can reset the inputenc settings like so: \documentclass{article} \usepackage[utf8]{inputenc} \usepackage[T2A,T1]{fontenc} \usepackage{CJK} \begin{document} {\fontencoding{T2A}\selectfont ф} \begin{CJK}{UTF8}{min} ル \end{CJK}% \sbox0{\input{utf8.def}}% {\... 8 Fixed in the inputenc shipped with LaTeX2e 2015/01/01 patch level 1. This isn't directly related to cm but looks like a bug in the core latex utf8 mapping, this fixes it for now \documentclass[varwidth]{standalone} \usepackage[utf8]{inputenc} \usepackage[T2A]{fontenc} \usepackage{xcolor} \makeatletter \AtBeginDocument{% \DeclareUnicodeCharacter{0403}{\@... 8 The more recent TeX distributions provide a Times-like font that supports the Cyrillic and Greek alphabets, loaded through the package tempora. \documentclass{disser} \usepackage[utf8]{inputenc} \usepackage[T1,T2A]{fontenc} \usepackage[english, russian]{babel} \usepackage[babel=true]{microtype} \usepackage{tempora} % Times for numbers in math mode \... 8 You're using the lmodern package, which sets the document fonts to Latin Modern, but it doesn't have support for cyrillic letters. There's Computer Modern Uncode for cyrillic and greek. For XeLaTeX, add \usepackage{fontspec} \setmainfont[ Ligatures=TeX, Extension=.otf, BoldFont=cmunbx, ItalicFont=cmunti, BoldItalicFont=cmunbi, ]{cmunrm} \... 8 In [russian]{babel} there is a mapping for cyrillic alphabetical numbering. Since the question asked for polyglossia I copied the definition of that mapping. Then you can redefine the subfigure counter to use the mapping. MWE: \documentclass[a4paper,oneside,12pt,serbianc]{article} % Serbian language and Cyrillic script \usepackage{fontspec} \... 7 This is a bug in babel, which appears when two cyrillic languages are chosen. I wrote about it in an article published in number 9 of ArsTeXnica, the journal of GuIT, the Italian TeX users group. You can get around the bug by correcting the behaviour of babel: \documentclass{article} \usepackage[T2A,T1]{fontenc} \usepackage[utf8]{inputenc} \usepackage[... Only top voted, non community-wiki answers of a minimum length are eligible
2019-10-16 08:07:48
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9588507413864136, "perplexity": 8864.868201652567}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570986666467.20/warc/CC-MAIN-20191016063833-20191016091333-00419.warc.gz"}
http://mymathforum.com/algebra/45478-completely-confused.html
My Math Forum Completely Confused Algebra Pre-Algebra and Basic Algebra Math Forum July 27th, 2014, 03:45 PM #1 Newbie   Joined: Jul 2014 From: Cohoes, NY Posts: 8 Thanks: 0 Math Focus: NONE!! Completely Confused I don't know how many of these things I can post in a day or even an hour. But this one has me completely confused. -9x(5x + 4) + 10(x + 3) Here is my calculation, which was completely wrong (still working with simplifying equations): -9x(5x + 4) + 10(x + 3) -45x + 36x + 10x + 30 19x + 30 Here was the answer the calculator gave me: -9x(5x + 4) + 10(x + 3) Reorder the terms: -9x(4 + 5x) + 10(x + 3) (4 * -9x + 5x * -9x) + 10(x + 3) (-36x + -45x^2) + 10(x + 3) Reorder the terms: -36x + -45x^2 + 10(3 + x) -36x + -45x^2 + (3 * 10 + x * 10) -36x + -45x^2 + (30 + 10x) Reorder the terms: 30 + -36x + 10x + -45x^2 Combine like terms: -36x + 10x = -26x 30 + -26x + -45x^2 Last edited by lwhitworth67; July 27th, 2014 at 04:00 PM. Reason: wrong calculation July 27th, 2014, 07:13 PM   #2 Senior Member Joined: Jul 2012 From: DFW Area Posts: 634 Thanks: 96 Math Focus: Electrical Engineering Applications Hi lwhitworth67, Quote: Here is my calculation, which was completely wrong (still working with simplifying equations): -9x(5x + 4) + 10(x + 3) -45x + 36x + 10x + 30 The first two terms should be -45x^2 - 36x. I think you will find that this will lead to the given answer. July 27th, 2014, 10:48 PM   #3 Senior Member Joined: Apr 2014 From: Europa Posts: 575 Thanks: 176 Quote: Originally Posted by lwhitworth67 -9x(5x + 4) + 10(x + 3) -45x + 36x + 10x + 30 19x + 30 $\displaystyle \ -9x(5x + 4) + 10(x + 3)= \\\;\\ = -45x^2 - 36x + 10x + 30= \\\;\\ = -45x^2 - 26x + 30 .$ Thread Tools Display Modes Linear Mode Similar Threads Thread Thread Starter Forum Replies Last Post Shamieh Computer Science 10 September 11th, 2013 09:09 PM l0velo Algebra 4 April 15th, 2011 01:28 AM mids1999 Applied Math 1 August 19th, 2009 11:06 AM malikah Algebra 5 September 11th, 2007 09:38 AM l0velo Linear Algebra 1 December 31st, 1969 04:00 PM Contact - Home - Forums - Cryptocurrency Forum - Top
2019-02-24 01:08:36
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5943233966827393, "perplexity": 4196.57783261699}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-09/segments/1550249569386.95/warc/CC-MAIN-20190224003630-20190224025630-00177.warc.gz"}
http://www.eng.tau.ac.il/~shaked/tech_rep.html
# Prof. Uri Shaked ### Technical Reports 1. U. Shaked, DIGNIS (Design for Ignorance in Non-Interacting Systems), User's Manual, The Department of Applied Mathematics, Weizmann Institute of Science, Internal Report, 1974. 2. U. Shaked, Quadratic Optimization of Partially Unknown Multivariable Feedback Systems, Technical Report, Weizmann Institute of Science, The Dept. of Applied Mathematics, 1974. 3. U. Shaked and N. Karcanias, The Use of Zeros and Zero Directions in Model Reduction, Cambridge University, Department of Engineering, CUED/F Control/TR99, 1975. 4. U. Shaked and B. Kouvaritakis, The Asymptotic Behavior of Root-Loci of Multivariable Systems, Cambridge University, Dept. of Engineering, CUED/F Control/TR100, 1975. 5. U. Shaked, Singular and Cheap Control: The Minimum-and the Nonminimum-Phase Cases, Technical Report of the National Research Institute for Mathematical Sciences of South Africa, TWISK 181, CSIR, 1980. 6. U. Shaked, Nearly Singular Optimal Linear Continuous Stationary Filtering, Technical Report of the Faculty of Engineering, Tel Aviv University, Tel Aviv, Israel, TAUEE/S-TR 4/82. 7. U. Shaked, Nearly Singular Finite Time Optimal Linear Filtering for Stationary Continuous Systems, Technical Report of The Faculty of Engineering, Tel-Aviv University, Tel- Aviv, Israel, TAUEE/S-TR 5/82. 8. U. Shaked, Nearly Singular Optimal Smoothing for Linear Continuous Stationary Processes, Technical Report of the Faculty of Engineering, Tel Aviv University, Tel Aviv, Israel, TAUEE/S-TR 6/82. 9. U. Shaked, A Simple Method for Deriving the Zeros of Linear Systems and Their Direction, Technical Report of the Faculty of Engineering Tel Aviv University, Tel Aviv, Israel, TAUEE/S-TR 4/84. 10. Y. Baram and U. Shaked, Order Minimality of Multivariable Continuous-Time Stochastic Realizations, EE Pub. No. 500, Faculty of Electrical Engineering, Technion, Israel Inst. of Technology, 1984. 11. E. Soroka and U. Shaked, The Robustness Properties of the LQG Optimal Regulator For Systems With Perfect Measurements, Technical Report of the Faculty of Engineering Tel-Aviv University, Tel Aviv, Israel, TAUEE/S-TR 506/84. 12. I. Yaesh and U. Shaked, Optimal Control and Estimation of Uncertain Linear Time Varying Systems, Technical Report of the Faculty of Engineering Tel-Aviv University, Tel Aviv, Israel, TAUEE/S-TR 26/91. 13. I. Yaesh and U. Shaked, Finite-Horizon Discrete-Time $H_{\infty}$ Filtering and Smoothing, Electrical Engineering-Systems Technical Reports, Tel Aviv University, EE-S-92-15, 1992. 14. Y. Theodor and U. Shaked, A Nash Game Approach to Discrete Time Mixed $H_{\infty}/H_2$ Filtering, Electrical Engineering Systems Technical Reports, Tel-Aviv University, EE-S-92-16, 1992. 15. Y. Theodor and U. Shaked, Game Theory Approach to $H_{\infty}$-Optimal Discrete-Time Fixed-Point and Fixed-Lag Smoothing, Electrical Engineering Systems Technical Reports, Tel-Aviv University, EE-S 2-21, 1992. 16. Y. Theodor and U. Shaked, $H_{\infty}$ Multiple Objective Robust Controllers for Infinite-Horizon Single Measurement Single Control Input Problems, Electrical Engineering-Systems Technical Reports, Tel-Aviv University, EE-S 93-7, 1993. 17. Y. Theodor and U. Shaked, A Game Theory Approach to Robust Discrete-Time $H_{\infty}$ Estimation, Electrical Engineering Systems Technical Reports, Tel-Aviv University, EE-S-93-8, 1993. 18. Y. Theodor and U. Shaked, Game Theory Approach to Mixed $H_{\infty}/H_2$ Filtering and Control, Electrical Engineering-Systems Technical Reports, Tel-Aviv University, EE-S-93-15, 1993. 19. Y. Theodor and U. Shaked, Robust Discrete-Time $H_{\infty}$ Filtering, Electrical Engineering-Systems Technical Reports, Tel-Aviv University, EE-S 93-16, 1993. 20. N. Berman and U. Shaked, $H_{\infty}$ Nonlinear Filtering. Electrical Engineering-Systems Technical Reports, Tel-Aviv University, EE-S-3-17, 1993. 21. U. Shaked and C. E. de Souza, Continuous-Time Tracking Problems in an $H_{\infty}$ Setting, Electrical Engineering-Systems Technical Reports, Tel-Aviv University, EE-S-93-18, 1993. 22. U. Shaked and N. Berman, $H_{\infty}$ Nonlinear Filtering of Discrete-Time Processes, Electrical Engineering-Systems Technical Reports, Tel-Aviv University, EE-S-93-20, 1993. 23. Y. Theodor and U. Shaked, Robust $H_{\infty}$ Control, Electrical Engineering-Systems Technical Reports, Tel-Aviv University, EE-S-4-2, 1994. 24. Y. Theodor, U. Shaked and N. Berman, Time Domain $H_{\infty}$ Identification, Electrical Engineering-Systems Technical Reports, Tel-Aviv University, EE-S-94-3, 1993. 25. M. Maroni, P. Bolzern, G. De Nicolao and , U.Shaked, Existence and Convergence of Solutions to the $H_{\infty}$ Sampled-Data Estimation Problem, Internal Report, n. 98-18. 26. E. Gershon, D. J. N. Limebeer, U. Shaked and I. Yaesh, Parametric BRLs for Discrete-Time Linear Systems with Polytopic Uncertainties , Tel-Aviv University, EE-S-99-4, 1999. % 27. E. Gershon, D. J. N. Limebeer, U. Shaked, I. Yaesh, Robust \hinf\ Tracking of Linear Discrete-Time Systems With Stochastic Uncertainties and Preview, Technical Report, Imperial College, EEE Dept., England, CP 99/1. 28. E. Gershon, D. J. Limebeer, U. Shaked and I Yaesh, Robust \hinf\ Filtering of Stationary Continuous-Time Linear Systems with Stochastic Uncertainties,Technical Report, Imperial College, EEE Dept., England, CP 99/2.
2022-06-26 19:31:51
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6402515172958374, "perplexity": 14229.905552074104}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656103271864.14/warc/CC-MAIN-20220626192142-20220626222142-00116.warc.gz"}
https://www.groundai.com/project/the-matter-gravity-entanglement-hypothesis/
The matter-gravity entanglement hypothesis # The matter-gravity entanglement hypothesis Bernard S. Kay Department of Mathematics, University of York, York YO10 5DD, UK 22email: bernard.kay@york.ac.uk ###### Abstract I outline some of my work and results (some dating back to 1998, some more recent) on my matter-gravity entanglement hypothesis, according to which the entropy of a closed quantum gravitational system is equal to the system’s matter-gravity entanglement entropy. The main arguments presented are: (1) that this hypothesis is capable of resolving what I call the second-law puzzle, i.e. the puzzle as to how the entropy increase of a closed system can be reconciled with the asssumption of unitary time-evolution; (2) that the black hole information loss puzzle may be regarded as a special case of this second law puzzle and that therefore the same resolution applies to it; (3) that the black hole thermal atmosphere puzzle (which I recall) can be resolved by adopting a radically different-from-usual description of quantum black hole equilibrium states, according to which they are total pure states, entangled between matter and gravity in such a way that the partial states of matter and gravity are each approximately thermal equilibrium states (at the Hawking temperature); (4) that the Susskind-Horowitz-Polchinski string-theoretic understanding of black hole entropy as the logarithm of the degeneracy of a long string (which is the weak string coupling limit of a black hole) cannot be quite correct but should be replaced by a modified understanding according to which it is the entanglement entropy between a long string and its stringy atmosphere, when in a total pure equilibrium state in a suitable box, which (in line with (3)) goes over, at strong-coupling, to a black hole in equilibrium with its thermal atmosphere. The modified understanding in (4) is based on a general result, which I also describe, which concerns the likely state of a quantum system when it is weakly coupled to an energy-bath and the total state is a random pure state with a given energy. This result generalizes Goldstein et al.’s ‘canonical typicality’ result to systems which are not necessarily small. journal: Foundations of Physics ## 1 The second law puzzle Let me begin my talk111This article is a written version of a talk given at the 18th UK and European Conference on Foundations of Physics (16-18 July 2016, LSE, London) by recalling one version of the second law of thermodynamics: The entropy of the universe begins low and increases monotonically. There are long-established and well-known arguments – see the discussion of ‘branch systems’ in Reichenbach () as also reviewed e.g. in Davies ()) – that other statements of the second law, in terms of what can and can’t happen with heat engines, refrigerators etc. follow from the above statement. As also explained in these references, the above statement leads to an explanation of time asymmetry; i.e. why, for example, it is commonplace to observe wine-glasses fall off tables and smash into pieces, but we never see lots of smashed pieces assemble themselves into wine-glasses and jump onto tables. But how do we define the entropy of a closed system? And why does it increase? A standard way of answering this (essentially due to Boltzmann around 1870) might be to consider for example what will happen if one starts with a system of gas molecules in the left half of a box (see Figure 2) and removes a partition, allowing the particles to diffuse into the right half of the box. In a classical discussion, one describes the states of this system with some given energy in terms of a dimensional phase space, the points of which are called ‘microstates’ and (see Figure 3) one imagines this phase space to be divided up into cells – called ‘macrostates’ – with the property that we cannot in practice distinguish between any pair of microstates in any single macrostate. One then defines the (‘coarse-grained’) entropy, , of a microstate by S=klogW (1) where is Boltzmann’s constant and is the volume of the macrostate containing that given microstate. The standard argument then is that (see Figure 3) the macrostate corresponding to “all the particles are in the left half of the box” will have a vastly smaller volume in phase space than the large macrostate which corresponds to “the molecules fill the box with roughly uniform density”. Hence, as time goes on and the state of the system wanders around the phase space accordingly, it is highly likely that the entropy – as defined by (1) will get bigger and stay bigger. However, this definition of entropy and this argument for its increase depends, unsatisfactorily, on the need to make judgments about what we can distinguish. For example, if (see again Figure 3) after previously ignoring such fine distinctions, we were to take the view that we can distinguish a state where, say, 48% of the particles are in the left half of the box and 52% in the right half from a state with roughly equal proportions222These numbers were not entirely randomly chosen, the talk being given shortly after the June 2016 Brexit referendum. then, at times for which the system’s microstate lies in the accordingly-defined new macrostate (obviously a subregion of the previously discussed large macrostate) then equation (1) would ascribe a different value to the entropy. Moreover, this unsatisfactory arbitrariness and vagueness in the definition of entropy is even more of a problem if we want to account for the version of the second law with which we began. For we are not even present to make any distinctions in the early universe! Turning to the quantum setting, von Neumann gave us long ago a quantum translation of Boltzmann’s equation (1). Given a description of our system in terms of a density operator, acting on the system’s Hilbert space , one defines its von Neumann entropy, , by SvN(ρ)=−ktr(ρlogρ). (2) But if we were to equate the physical entropy, , with and if satisfies the usual unitary time evolution rule ρ(t)=U(t)ρ(0)U(t)−1 then we would conclude that Sphysical(ρ(t))=constant. in contradiction with the second law. We shall call this the second law puzzle. One can overcome this difficulty by defining quantum counterparts to the above classical coarse-graining, but of course one then would have the same unsatisfactory vagueness and subjectivity as we discussed above in the classical case. More interestingly, one can seek to exploit a feature of quantum mechanics which has no classical counterpart: If we have a pure state, described by a density operator, , which is a projector onto a vector, , in a Hilbert space, , which arises as the tensor product, Htotal=HA⊗HB of two Hilbert spaces, and , then the reduced density operator, on , defined as the partial trace, , of over , will typically have . We remark that • This partial trace is characterized by the property that, if is a (self-adjoint) operator on , then tr(ρAO)HA=⟨Ψ(O⊗I)|Ψ⟩Htotal. • Both reduced density operators have equal von Neumann entropies: SvN(ρA)=SvN(ρB) (3) and this common value is often known as the A-B entanglement entropy of the total state-vector . In a variant of the ‘environment paradigm for decoherence’ or, from another point of view, a variant of a possible approach to quantum statistical mechanics, this formalism is often applied in the case that A is interpreted as standing for some ‘system’ and B for the system’s ‘environment’ or ‘energy bath’ and is then interpreted as the entropy of the system due to entanglement with the environment. So the environment paradigm gives us an objective notion of entropy. However, there remain problems: • It only offers a notion of entropy for open systems. • There are lots of ways of decomposing a given as . How we choose to decompose it depends on subjective choices and, again, we are not around in the early universe to make those choices. What I’d like to point out is that one can envisage an alternative physical use of this mathematical fact: Suppose there’s some decomposition that’s physically natural, then maybe we could define the entropy of a total closed system by Stotal=SvN(ρA)(=SvN(ρB))(= A-B entanglement entropy) (4) rather than interpreting this mathematical quantity as the entropy of the A-subsystem! We propose that the identification: is the right choice. This is our matter-gravity entanglement hypothesis. (See Kay1998a (), Kay1998b () and KayAbyaneh () for early papers, and KayEntropy () and the remainder of the present article for recent partial overviews and further references.) In support of this, we note that the decomposition has to be meaningful throughout the entire history of the universe: E.g. we couldn’t identify A with photons and B with nuclei+electrons because these notions are not even meaningful until the photon epoch. We content ourselves, though, with going back to just after the Planck epoch; we assume that a low-energy quantum gravity theory holds there and throughout the entire subsequent history of the universe and that this is a conventional (unitary) quantum theory with . We will also assume that the initial degree of matter-gravity entanglement is low. (We leave it for a future theory of the pre-Planck era to explain that.) These assumptions then appear to be capable of offering an explanation of the second law in the form stated at the outset since one can argue that an initial state with a low degree of matter-gravity entanglement will, because of matter-gravity interaction, get more entangled, plausibly monotonically, as time increases. At least the question of whether the second law holds becomes a question which, in principle, can be answered mathematically once we specify the (low-energy) quantum gravity Hamiltonian (i.e. the generator of the unitary time-evolution) and the initial state. What we have called the second law puzzle would then be resolved because once we define entropy as matter-gravity entanglement entropy (rather than as the von Neumann entropy of the total state) there is no conflict between its increase and a unitary time-evolution. ## 2 The information loss puzzle (Hawking 1976) The celebrated result of Hawking HawkingCMP75 () is that a black hole formed by the dynamical collapse of a star will emit thermal radiation at the Hawking temperature, given, in the case of a spherically symmetric electrically neutral black hole by kTHawking=18πGM (5) where is the black hole mass (and we take ). As Hawking explained in that work, one expects that such a radiating black hole will lose mass, increasing further its temperature, and eventually evaporate. During this whole process of collapse to a black hole and subsequent evaporation, one expects the entropy of the total system to increase monotonically.333Without wishing to imply that they are necessarily exactly additive, we note that while the entropy of the black hole (given by (6)) will decrease because the horizon area will decrease, one expects that this will be more than compensated by the increased entropy of the sphere of emitted Hawking radiation which is growing in size at the speed of light and within which, moreover, the later radiation will be hotter than that emitted earlier. The version of the information loss puzzle HawkingInfoLoss () that I shall adopt here is the puzzle as to how this entropy increase can be reconciled with an assumption of unitary time evolution. Stated in this way, I think it is clear that the information loss puzzle is nothing but a special case of our Second Law Puzzle; we recall here that this is the puzzle that, if one equates with ), then must be constant. I suggested in Kay1998a () and Kay1998b () that the resolution to the information loss puzzle is simply the special case of the above proposed resolution to the second law puzzle. Namely, is not . Rather is the total state’s matter-gravity entanglement entropy. As I already said in the more general context in Section 1, this is not a unitary invariant and – it is reasonable to assume – would increase, thus offering to resolve the puzzle. That it also offers this resolution to the information loss puzzle lends, is, in my view, further evidence that our matter-gravity entanglement hypothesis is on the right track. ## 3 The thermal atmosphere puzzle A black hole in a box in equilibrium with its thermal atmosphere (see Figure 5) is traditionally taken to be in a total Gibbs state (in particular a total mixed state) at the Hawking temperature Everyone agrees that the entropy of this system has (at least up to small corrections) the value SHawking=4πkGM2=kA/4G. (6) where is the surface area of the event horizon (). The thermal atmopshere puzzle Page () WaldLivRev () is that one can give seemingly convincing arguments for each of the following three, at first sight seemingly mutually contradictory, statements about the nature and origin of this entropy: • It is the entropy of the gravitational field (so mostly ‘residing’ in the black hole). • It is the entropy of the thermal atmosphere (so apart from the graviton component, consisting mainly of matter). • It is the sum of the above two entropies. Our proposed resolution of the puzzle begins by postulating that it is not actually the case that the total state is a Gibbs state; rather, we propose, the total state is pure, but entangled between gravity ( the black hole) and matter ( its atmosphere) in such a way that each are approximately Gibbs states (at the Hawking temperature). We further suggest, in line with our matter-gravity entanglement hypothesis, that is really this state’s matter-gravity entanglement entropy. This offers to resolve the puzzle in the following way: The first entropy can be regarded, according to the environment paradigm, as the entropy of the open system consisting of the gravitational field due to its matter environment; the second the entropy of the open system consisting of the matter due to its gravity environment. But, by (3), these are actually equal and so, in this environment-paradigm sense, both statements are therefore true, without contradiction. On the other hand, there is no reason why the third statement should be true in any sense and in fact, on our hypothesis it is clearly not true – the total entropy being, by (4) not the sum of the first two, but rather, equal to each of them. The fact that it seems capable of providing this resolution to the thermal atmosphere puzzle provides further support for the validity of our matter-gravity entanglement hypothesis. ## 4 The weak string-coupling limit of black-hole equilibrium states and black hole entropy Some of the most interesting work towards computing (in certain cases) or, at least, gaining a better understanding of, black hole entropy has been within string theory. Here I shall briefly recall the basic idea due to Susskind Susskind () and one particular line of development by Horowitz and Polchinski HoroPolch () Horowitz () which leads to an explanation of how the entropy of spherically symmetric black holes scales with (the square of the black-hole mass), albeit the argument is semi-qualitative and does not tell us the constant term (so doesn’t explain the factor of in (6)). First I will outline the Susskind-Horowitz-Polchinski (SHP) argument. Then I will criticize it. Then I will propose a modification of the SHP argument which is free from the criticisms I raise and is consistent with the understanding of black-hole equilibrium states on the matter-gravity entanglement hypothesis that I outlined in Section 3. The SHP argument HoroPolch () Horowitz () is in two steps444We adopt similar simplifications to those adopted in HoroPolch () Horowitz (). Thus the spacetime dimension is taken to be 4 and the power-law prefactors in the densities of states are ignored. See however KayMore () for the importance of those prefactors in my proposed modification of the SHP argument.: First (see Figure 6) one argues that, as one scales the string coupling-constant, , down and the string length, up, keeping Newton’s constant fixed, a black hole goes over to a long string. This will have density of states (i.e. number of states per unit energy, where we use to denote energy) approximately of the form of a constant times . Secondly, one equates the entropy, , with “ at constant times when constant times whereupon constant times . Our criticism of this is that it is not correct to equate an entropy with the logarithm of a density of states. (Nor indeed, in other string theory work, with the logarithm of a degeneracy – see KayMod () and KayEntropy ().) Indeed it only ever makes sense in physics to take the logarithm of a dimensionless quantity but a density of states has of course the dimensions of inverse energy! Our proposed modification of the SHP scenario KayMod () KayMore () is to consider, in place of the limit {black hole}→{long string}, the limit {black hole in equilibrium with thermal atmosphere in a box}→ {long string in equilibrium with atmosphere of small strings in a % suitably rescaled box}. The key fact HoroPolch () Horowitz () about a string equilibrium state of this latter type is that (in a certain approximation where we ignore certain power-law prefactors – see Footnote 4) the long string and its stringy atmosphere will have densities of states of the exponential form: σlongstring(ϵ)∼ceℓsϵ,σstringyatmosphere(ϵ)∼c′eℓsϵ (7) where the constants and may be different, but, importantly the exponents are the same. I have demonstrated (see Section 5 for a discussion of the proof) that: Theorem 1: For any pair of weakly coupled systems (to be called here ‘system’ and ‘bath’) with densities of states as in (7) a randomly chosen pure equilibrium state with total energy will, with very high probability, have a system-bath entanglement entropy approximately equal to . It will also be such that the reduced states of system and bath separately each have energy and are each approximately thermal at temperature Applying this theorem and reading ‘long string’ for ‘system’ and ‘stringy atmosphere‘ for ‘bath’ (or vice versa) and equating the black hole mass, , with a constant times and the entanglement entropy of this theorem with the matter-gravity entanglement entropy of the black hole equilibrium state at constant times (as in the unmodified argument) the latter entropy will thus be a constant times . Thus we achieve a corrected string explanation of this formula for the black hole entropy which is not subject to the criticism we made of the original SHP approach. Moreoever making the same substitution, constant times , the temperature formula for the reduced states of the long string and of its stringy atmosphere goes over to the temperature formula a constant times , which agrees with the Hawking temperature formula (5) (up to a constant).555 Intriguingly, as pointed out in KayMod (), if one equates with and equates with , then one gets the right value both for the Hawking temperature and the Hawking entropy. However, as explained in KayMod () and KayMore () this numerical coincidence should be interpreted with caution. That ends my discussion of my matter-gravity entanglement hypothesis and of how it offers a resolution to the three puzzles: the second law puzzle, the black hole information loss puzzle, and the thermal atmosphere puzzle and, finally, in this section, of how it enables a modification of the SHP string approach to black hole entropy which is free from the criticism666To provide further perspective on that criticism, let us recall that the attempt to provide a microscopic explanation of thermodynamical behaviour in terms of a classical statistical mechanics has often been criticized because it requires the introduction of an ad hoc quantity with the dimensions of action in order to provide a unit of volume in phase space. It has been said that this shortcoming of classical statistical mechanics is overcome in quantum statistical mechanics where a suitable power of the quantity effectively provides the right volume element. One might re-express the main thesis of this section by saying that, in a similar way, the need to introduce an ad hoc dimensionful quantity as in the SHP approach to black hole entropy and the resolution of that difficulty along the lines explained in the main text indicates that, to have a satisfactory microscopic explanation of thermodynamical behaviour, a quantum statistical mechanics is also insufficient and what is needed, instead, is a quantum-gravitational statistical mechanics based on our matter-gravity entanglement hypothesis. which I made of the original SHP approach. In the remainder of the talk I would like to supply some of the details about how I proved the above theorem. ## 5 Explanations of thermality: traditional and modern Theorem 1 in fact relies on a general theorem – which is stated below as Theorem 2 – which I obtained KayThermality () in a general setting where one has a total system (in KayThermality () I abbreviate this with the the term ‘totem’ and I shall follow that terminology here) consisting of a (quantum) system weakly coupled to an energy bath. Such a totem will have a Hamiltonian of form H=Hsystem+Hbath+Hinteraction on Hsystem⊗Hbath where is assumed to be sufficiently weak that it can be ignored for the purposes of counting energy levels; and each have positively supported, locally finite, discrete spectrum with monotonically increasing densities of states, σsystem(ϵ)andσbath(ϵ). Theorem 2 may be considered to generalize a result of Goldstein, Lebowitz, Tumulka and Zanghi (GLTZ) GoldsteinEtAl () (see also PopescuEtAl ()) which explains why it is that a small system in contact with a large energy bath will typically be in an (approximate) thermal equilibrium state. So I will first briefly recall that result: ### 5.1 Thermality in the case the system is small The GLTZ explanation is itself a modern replacement for the earlier traditional explanation of the thermality of a small system in contact with a heat bath, so let me recall that first. The traditional explanation is based on a mathematical theorem which tells us that if the totem is in a microcanonical ensemble with energy in a narrow band around some total energy , then the small system will be approximately in a thermal equilibrium state with temperature, given by the formula (note that the dimensionful argument of the logarithm is innocuous here because the logarithm is differentiated): 1kTsystem=ddϵlogσbath(ϵ)|ϵ=E. (8) The modern explanation GoldsteinEtAl () is based on a mathematical theorem (proven in GoldsteinEtAl ()) that if the totem state is a pure state, randomly chosen from the set of all pure states with totem energy in a narrow band around (where the random choice is with respect to a natural measure on the set of all these pure states) then the small system will very probably be very close to the same thermal equilibrium state with a temperature given by the same formula (8). The advantage of the “modern” over the “traditional” point of view is that it bases a theory of how systems get themselves into (approximate) Gibbs states on the same foundational assumption that we usually make for the foundations of quantum mechanics – namely that the total state of a full closed system is a pure (vector) state. ### 5.2 What happens when System and Energy Bath are of comparable size? One might think that one could apply the GLTZ result directly to the case our totem is the string equilibrium state illustrated in Figure 7, identifying, say, the long string with our ‘system’ and the stringy atmosphere with our ‘energy bath’. However, neither of these can be regarded as small with respect to the other. Here we should clarify that ‘small’ in this context would mean having much more widely spaced energy levels, i.e. having a much lower density of states. Instead both densities of states are (ignoring the power-law prefactors I mentioned earlier) of the exponentially increasing form (7). It turns out in general, that when the system and the energy bath are of comparable size, then – on both the traditional assumption of a totem microcanonical ensemble and the modern assumption of a random total pure state with energy in a small band – it is no longer necessarily the case that either system or energy bath will probably be in a thermal equilibrium state. However, I have shown KayThermality () with regard to the modern approach: Theorem 2 There is a special density operator (see the Appendix for details) ρmodapproxsystem  on  Hsystem (9) such that, given a random vector, , with energy in a narrow band around , then the partial trace of over is very probably very close to . (And similarly with system energy bath). But it is important to realize that when system and energy bath are of comparable size, is not always thermal. (And neither, by the way, is the reduced state of the system thermal when the total state is in a traditional microcanonical ensemble.) E.g. if and take, respectively, the power law forms , (the typical behaviour of ordinary matter when and are comparable in size to Avogadro’s number) then the system ‘energy probability density’, KayThermality () will be a Gaussian (in fact the same Gaussian on both traditional and modern assumptions) rather than the Gibbsian distribution characteristic of a thermal state. See Figures 10 and 11. ### 5.3 The special nature of exponential densities of states However, it is shown in KayThermality (), regarding the modern approach777A similar result to Theorem 3 holds for the traditional (microcanonical) approach, except that (now neglecting logarithmic terms) in place of one finds KayThermality () that the system and the energy bath have entropy . The difference between these two results is interesting since it demonstrates that, in general, the traditional and modern approaches don’t give the same results. (It is also interesting since the “right value for the Hawking entropy” mentioned in Footnote 5 depends on the denominator being 4 – rather than 2). Theorem 3: When system and energy-bath densities of states both take the exponential form of Equation (7): • and are (close to888See KayThermality () for the sense in which these states are close to thermal.) thermal at temperature . (And each have mean energy .) • Also, the system-energy bath entanglement entropy, , is approximately . 999The exact result (KayThermality, , Endnote 29) is . Theorem 1 of Section 4 clearly follows immediately from Theorems 2 and 3. ## Appendix: Details on ρmodapproxsystem In this appendix we give the detailed formula for the special density operator (9). Define the (-dimensional) Hilbert space, ( assumed large) consisting of elements with total energy in a narrow band to be the closed span of eigenstates of the total Hamiltonian with energies, . For convenience, replace the system of interest by a system with equally spaced energy levels with spacing equal to – each energy level, , having degeneracy, (so that the new system will have the same density of states, as the original system). (Note that then .) We note first that the traditional microcanonical density operator, is then easily seen to have reduced density operator on equal to ρmicrocsystem=M−1E∑ϵ=Δnbath(E−ϵ)nsystem(ϵ)∑i=1|ϵ,i⟩⟨ϵ,i| where denotes a basis for the -dimensional degeneracy subspace of with energy (assumed to be a multiple of ) and the sum over is over multiples of . The modern replacement for this result is that a random pure density operator, , on will have a reduced density operator on which (as is argued in KayThermality ()) is very probably very close to where ρmodapproxsystem=M−1  times Ec∑ϵ=Δnbath(E−ϵ)nsystem(ϵ)∑i=1|ϵ,i⟩⟨ϵ,i|+E∑ϵ=Ec+Δnsystem(ϵ)nbath(E−ϵ)∑i=1|˜ϵ,i⟩⟨˜ϵ,i| where is the energy at which and the span an orthonormal basis of an -dimensional subspace of the -dimensional) energy- subspace of which depends on in a way explained in detail in KayThermality (). ## Afterword To end, let me mention some related aspects of the matter-gravity entanglement hypothesis that we have not had time to discuss. One is an extension of the theory beyond closed systems to include open systems. For this, we refer to (KayAbyaneh, , Endnote (xii)) or KayEntropy (). Another concerns the relevance of the matter-gravity entanglement hypothesis to the measurement problem in quantum mechanics and a possible resolution to the Schrödinger Cat puzzle. For this, see Kay1998a (), Kay1998b () and KayAbyaneh (). Finally, the papers KayInstab (), KayLupo () (see also KayEntropy () for a brief outline of this work) include a discussion of a possible mechanism whereby, when one passes from a quantum field theory in curved spacetime description to a description in which the backreaction of the stress-energy tensor on the metric is taken into account, the horizon of an enclosed (say Kruskal) black hole becomes unstable with the consequence that entanglement between the right and left Kruskal wedges in a quantum theory in curved spacetime context transmutes into entanglement between matter and gravity – in support of the solution to the thermal atmosphere puzzle presented in Section 3. ## References • (1) Reichenbach, H.: The Direction of Time. University of California Press, Berkeley (1971) • (2) Davies, P.C.W.: The Physics of Time Asymmetry. University of California Press, Berkeley and Los Angeles (1977) • (3) Kay, B.S.: Entropy defined, entropy increase and decoherence understood, and some black-hole puzzles solved. arXiv:hep-th/9802172 (1998) • (4) Kay, B.S.: Decoherence of macroscopic closed systems within Newtonian quantum gravity. Class. Quant. Grav. 15, L89-L98 (1998) [hep-th/9810077] • (5) Kay, B.S., Abyaneh, V.: Expectation values, experimental predictions, events and entropy in quantum gravitationally decohered quantum mechanics. arXiv:0710.0992 (2007) • (6) Kay, B.S.: Entropy and quantum gravity. Entropy 17, 8174 (2015) [arXiv:1504.00882] • (7) Hawking, S.W.: Particle creation by black holes. Commun. Math. Phys. 43, 199-220 (1975) • (8) Hawking, S.W.: Breakdown of predictability in gravitational collapse. Phys. Rev. D 14, 2460–2473 (1976) • (9) Page, D.: 2005 Hawking radiation and black hole thermodynamics. New J. Phys. 7, 203 (2005) [hep-th/0409024] • (10) Wald, R.M.: The thermodynamics of black holes. Living reviews in relativity 4, no. 1 (2001). • (11) Susskind, L.: Some speculations about black hole entropy in string theory. arXiv:hep-th/9309145 (1993) • (12) Horowitz, G., Polchinski, J.: A correspondence principle for black holes and strings. Phys. Rev. D 55, 6189-6197 (1997) • (13) Horowitz, G.: Quantum states of black holes. In: Wald, R.M. (ed.) Black Holes and Relativistic Stars. University of Chicago Press, Chicago (1998) [arXiv:gr-qc/9704072] • (14) Kay, B.S.: More about the stringy limit of black hole equilibria. arXiv:1209.5110 (2012) • (15) Kay, B.S.: Modern foundations for thermodynamics and the stringy limit of black hole equilibria. arXiv:1209.5110 (2012) • (16) Kay, B.S.: On the origin of thermality. arXiv:1209.5215 (2012) • (17) Goldstein, S., Lebowitz, J.L., Tumulka, R., Zanghi, N.: Phys. Rev. Lett. 96, 050403 (2006) [cond-mat/0511091] • (18) Popescu, S., Short. A.J., Winter, A.: The foundations of statistical mechanics from entanglement: Individual states vs. averages. Nature Physics 2, 754 (2006) [quant-ph/0511225] • (19) Kay, B.S.: Instability of enclosed horizons. General Relativity and Gravitation 47, 31 (2015) [arXiv:1310.7395] • (20) Kay, B.S., Lupo, U.: Non-existence of isometry-invariant Hadamard states for a Kruskal black hole in a box and for massless fields on 1+1 Minkowski spacetime with a uniformly accelerating mirror. Class. Quantum Grav. 33 215001 (2016) [arXiv:1502.06582] You are adding the first comment! How to quickly get a good reply: • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made. • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements. • Your comment should inspire ideas to flow and help the author improves the paper. The better we are at sharing our knowledge with each other, the faster we move forward. The feedback must be of minimum 40 characters and the title a minimum of 5 characters
2021-01-25 08:17:50
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8565700054168701, "perplexity": 775.4898045456448}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610703565376.63/warc/CC-MAIN-20210125061144-20210125091144-00287.warc.gz"}
http://accessemergencymedicine.mhmedical.com/content.aspx?bookid=431&sectionid=40216667
Chapter 45 Umbilical vessel catheterization can be used as a reliable method of obtaining rapid vascular access in the neonate. Umbilical vessel catheters may be used for fluid resuscitation, blood transfusion, medication administration, frequent blood sampling, and cardiovascular monitoring.1–4 Either the umbilical artery or vein may be used for vascular access. The artery can usually be accessed within the first 24 hours of life. It is occasionally possible to use the umbilical artery up to 7 days after birth.1 The umbilical vein can be accessed for up to 2 weeks of age.1,5 Umbilical artery catheterization is more desirable than umbilical vein catheterization because it allows frequent arterial blood gas sampling and continuous blood pressure monitoring, in addition to fluid, blood, and medication administration. Unfortunately, umbilical artery catheterization is more difficult and time consuming to perform, especially in unskilled hands. Therefore, umbilical vein catheterization is the preferred procedure for the infant in shock and in need of rapid resuscitation. Arterial access can be obtained later in a more controlled environment, such as in the neonatal intensive care unit. Umbilical vessel catheterization can lead to serious complications and should be reserved for the patient in whom peripheral venous access attempts have been unsuccessful.4,6 The fetal circulatory system is quite different from that of the neonate or infant (Figure 45-1). Oxygenated blood from the placenta travels via the umbilical vein, through the ductus venosus in the liver, to the inferior vena cava, and into the right atrium. Oxygenated blood from the inferior vena cava preferentially enters the left atrium through the foramen ovale. It then enters the left ventricle, then the aorta. This oxygen-rich blood supplies the brain prior to mixing with the oxygen-poor blood coming through the ductus arteriosus. Deoxygenated blood from the superior vena cava enters the right ventricle and is pumped to the pulmonary artery. It then passes through the ductus arteriosus to meet the oxygenated blood in the aorta. ###### FIGURE 45-1 The fetal circulation. Pulmonary vascular resistance decreases dramatically as the infant takes its first breaths. The systemic vascular resistance increases when the umbilical cord is clamped. The foramen ovale closes with the combination of decreased pulmonary artery pressure and increased systemic resistance. The ductus arteriosus closes within 24 to 48 hours due to the release of prostaglandins and increased oxygen tension. The ductus venosus closes when the umbilical cord is clamped. The umbilical vein and arteries can easily be differentiated by examination of a cross section of the umbilical cord (Figure 45-2). The umbilical vein is a single vessel with thin walls and a large lumen. It is usually flattened in one direction. There are two thick-walled umbilical arteries that are significantly smaller in diameter than the umbilical vein. Occasionally, only a single umbilical artery is present. ... Sign in to your MyAccess profile while you are actively authenticated on this site via your institution (you will be able to verify this by looking at the top right corner of the screen - if you see your institution's name, you are authenticated). Once logged in to your MyAccess profile, you will be able to access your institution's subscription for 90 days from any location. You must be logged in while authenticated at least once every 90 days to maintain this remote access. Ok If your institution subscribes to this resource, and you don't have a MyAccess profile, please contact your library's reference desk for information on how to gain access to this resource from off-campus. ## Subscription Options ### AccessEmergency Medicine Full Site: One-Year Subscription Connect to the full suite of AccessEmergency Medicine content and resources including advanced 8th edition chapters of Tintinalli’s, high-quality procedural videos and images, interactive board review, an integrated drug database, and more.
2017-03-25 19:44:20
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.26239871978759766, "perplexity": 5324.382029344716}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-13/segments/1490218189032.76/warc/CC-MAIN-20170322212949-00447-ip-10-233-31-227.ec2.internal.warc.gz"}
https://www.albert.io/learn/ap-physics-c-mechanics/question/change-in-momentum-in-a-2d-rebound
Limited access A ball of mass $m$ moving with a velocity of magnitude $v_0$ at an angle $\theta$ to the $x$-axis strikes a wall and bounces with the same momentum and angle $\theta$, as shown in the figure above. What is the magnitude of the change in momentum of the ball? A $\text{Zero}$. B $2mv_0 \hat i$ C $2mv_0 \sin\theta \hat i$ D $2mv_0 \cos\theta\hat i$ E $2mv_0 \cos\theta\hat j$ Select an assignment template
2017-04-29 03:26:42
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3041456341743469, "perplexity": 145.73896667362573}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917123270.78/warc/CC-MAIN-20170423031203-00244-ip-10-145-167-34.ec2.internal.warc.gz"}
https://www.gradesaver.com/textbooks/math/algebra/elementary-algebra/chapter-6-factoring-solving-equations-and-problem-solving-6-1-factoring-by-using-the-distributive-property-problem-set-6-1-page-243/80
## Elementary Algebra $x = 0, -\frac{7}{5}$ 1. Rewrite the equation so that it is set equal to zero: $7x = - 5x^{2}$ $7x + 5x^{2} = 0$ 2. Factor out $x$ from the equation, set each factor equal to zero, and solve: $x(7 + 5x) = 0$ $x=0$ $7 + 5x = 0$ $5x = -7$ $x = -\frac{7}{5}$ $x = 0, -\frac{7}{5}$
2018-08-16 21:01:24
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9754735827445984, "perplexity": 153.6880663955301}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-34/segments/1534221211167.1/warc/CC-MAIN-20180816191550-20180816211550-00491.warc.gz"}
http://tex.stackexchange.com/questions/171891/how-to-create-a-format-file-with-initex-to-speed-up-processing-time
# How to create a format file with initex to speed up processing time? I have a file called matteo.tex which contains lots of macros. I'd like to create a format file with initex so that, every time tex processes a file that contains some of matteo.tex macros, it doesn't have to waste time processing the file matteo.tex. How to do it? And, then, how to make tex load the format generated from matteo.tex? (matteo.tex is all written in plain tex). - I've found the solution by myself: have a look at TEX BY TOPIC, A TEXNICIAN’S REFERENCE by VICTOR EIJKHOUT at page 257. –  Matteo Apr 17 '14 at 17:34 Please make this a proper answer. –  Martin Schröder Apr 17 '14 at 19:32 @MartinSchröder What do you mean? Can I answer my own question? –  Matteo Apr 18 '14 at 12:08 Yes. –  Martin Schröder Apr 18 '14 at 13:52 To make a format file from matteo.tex, you should simply add at the end of matteo.tex the control word \dump and then type on the command line tex matteo Now a file called matteo.fmt should have been generated. To load such file, you should type on the command line tex &matteo. If the format in matteo.tex does not extends plain tex, in order to create matteo.fmt you can start tex with the option -ini, i.e. you start initex. - Thank you for this question and answer! Can you tell if the speed-up is noticeable? –  morbusg Apr 19 '14 at 9:46 @morbusg I made up a test, running 50 consecutive times the compilation in both ways. It resulted that the usual way took about 9 seconds, and the ".fmt" way took about 7.5 seconds: so the difference is not so noticeable. But I think that if there were more macros in matteo.tex, the difference in time would be greater. Ah I forgot to say that in Unix systems, you have to write tex \&matteo instead of tex &matteo. –  Matteo Apr 19 '14 at 12:21
2015-04-19 23:00:57
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9603631496429443, "perplexity": 2293.9440108841936}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-18/segments/1429246642012.28/warc/CC-MAIN-20150417045722-00140-ip-10-235-10-82.ec2.internal.warc.gz"}
http://mathhelpforum.com/algebra/124395-sat-problem.html
# Math Help - SAT Problem 1. ## SAT Problem If g(x) = x * 2^x, then g(a+1) - g(a)= (a) (a+2)2^a (b) (2a+1)2^a (c) (2a-1)2^1 (d) (a+1)2^a+1 (e) (a)2^a+1 2. Originally Posted by juliak If g(x) = x * 2^x, then g(a+1) - g(a)= (a) (a+2)2^a (b) (2a+1)2^a (c) (2a-1)2^1 (d) (a+1)2^a+1 (e) (a)2^a+1 Try some values for a and see which cases are possible. For example if a=0 g(a+1)-g(a)=g(1)-g(0)=2 and (a)=2, (b)=1, (c)=-2, (d)=2, (e)=1 so we have elliminated all but (a) and (d), now try another value. CB 3. Thank you very very much! (sorry but) another question I have: Which of the following CANNOT represent the degree measure of an equiangular polygon? (a) 165 (b) 162 (c) 140 (d) 125 (e) 90 4. Originally Posted by juliak Thank you very very much! (sorry but) another question I have: Which of the following CANNOT represent the degree measure of an equiangular polygon? (a) 165 (b) 162 (c) 140 (d) 125 (e) 90 The internal sum of a polygon is given by $(n - 2)180^\circ$, where $n$ is the number of sides. If it is a regular polygon then all the angles are equal. Since you have $n$ angles, then if $\theta$ is one of the angles $n\theta = (n - 2)180^\circ$. So $\theta = \frac{(n - 2)180^\circ}{n}$. So now substitute some values for $n$ and see which ones that you have listed apply.
2015-04-25 07:21:40
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 7, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8202327489852905, "perplexity": 3113.329577974314}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-18/segments/1429246647589.15/warc/CC-MAIN-20150417045727-00170-ip-10-235-10-82.ec2.internal.warc.gz"}
https://gottwurfelt.com/2016/01/21/1889/
# Snow day? The National Weather Service’s Weather Prediction Center now has probabilistic winter precipitation guidance.  You can get maps showing • the $n$th percentile of the predicted distribution of accumulation for $n = 5, 10, 25, 50, 75, 90, 95$; • dually to that, contours on which the 10th, 20th, …, 90th percentile of the predicted distribution is a selected constant for the next 72 hours, or for 24- or 48-hour subsets of that time period.  Here’s a more in-depth description. Looks like the mid-Atlantic is in for a rough ride. Stay home. Have some French toast.  If you do go outside, ignore the wind chill.
2022-09-30 23:19:11
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 2, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7134740352630615, "perplexity": 4659.658785184269}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335504.37/warc/CC-MAIN-20220930212504-20221001002504-00306.warc.gz"}
https://space.stackexchange.com/questions/39448/what-is-esas-solar-orbiters-target-inclination
# What is ESA's Solar Orbiter's target inclination? The BBC News article European SolO probe ready to take on audacious mission links to the ESA video shown below and contains this 5 minute audio interview with NASA deputy project scientist on the mission Holly Gilbert who outlines several interesting aspects of the mission and complementary measurements with Parker Solar Probe. She mentions somewhere (I don't have the time code) that the Solar Orbiter will have a view of the Sun's poles, and the video shows what looks like an inclined elliptical orbit. What inclination are they targeting? I couldn't find any information about that on the ESA page https://sci.esa.int/web/solar-orbiter/ and this ESA Red Book is from 2011, so it may not contain up-to-date information. The inclination will be adjusted a few times during the mission: The nominal mission of seven years will see a maximum orbital inclination relative to the solar equator of 25°. During the extended mission, additional Venus GAMs could allow the orbital inclination to increase to about 34°. The main scientific activity will take place during the near-Sun encounter and high-latitude parts of each orbit, with different science goals envisioned for each orbit. From Solar Orbiter: Exploring the Sun-heliosphere connection (ArXiv, Researchgate, and published in Solar Physics) Figure 3 shows the solar latitude of the spacecraft versus time, and the extrema reflect the orbit's inclination for a January 2017 launch which didn't happen. While the orbit will be different, this at least illustrates how the inclination was planned to slowly increase over time after each orbit-lowering flyby. (For example, this paywalled paper lists the launch date as Jan-2017 (Mar-2017 and Sep-2018 back-ups)) after flyby approx. date approx. inc ----------- ------------ ----------- Venus 1 06-2017 7 Earth 1 04-2018 14 Earth 2 + Venus 2 06-2020 17 Venus 3 04-2022 25 Venus 4 01-2024 31 Venus 5 04-2025 34 Figure 3 Mission profile for a January 2017 launch, showing heliocentric distance (top) and latitude (bottom) of Solar Orbiter as a function of time. Also indicated are the times at which gravity assist maneuvers at Venus and Earth occur (blue).
2021-12-08 16:03:01
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2971837818622589, "perplexity": 4919.023633723757}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964363515.28/warc/CC-MAIN-20211208144647-20211208174647-00512.warc.gz"}
https://solarforecastarbiter-core.readthedocs.io/en/latest/generated/solarforecastarbiter.metrics.deterministic.over.html
# solarforecastarbiter.metrics.deterministic.over¶ solarforecastarbiter.metrics.deterministic.over(obs, fx)[source] The OVER metric. $\text{OVER} = \int_{p_\min}^{p_\max} D_n^*(p) dp$ where: $\begin{split}D_n^* = \begin{cases} (D_n - V_c) & D_n > V_c \\ D_n^* = 0 & \text{otherwise} \end{cases}\end{split}$ with D_n and V_c defined the same as in KSI. Parameters: obs ((n,) array-like) – Observed values. fx ((n,) array-like) – Forecasted values. over (float) – The OVER metric of the forecast. Notes The calculation of the empirical CDF uses a right endpoint rule (the default of the statsmodels ECDF function). For example, if the data is [1.0, 2.0], then ECDF output is 0.5 for any input less than 1.0.
2020-08-08 03:41:19
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7040722370147705, "perplexity": 7381.253252413743}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439737238.53/warc/CC-MAIN-20200808021257-20200808051257-00257.warc.gz"}
https://www.physicsforums.com/threads/electric-field-in-the-center-of-an-arc.920866/
# Electric field in the center of an arc 1. Jul 23, 2017 ### SetepenSeth 1. The problem statement, all variables and given/known data Consider a conductor wire with a charge Q uniformly distributed, shaped in the form of an arc of radius R and amplitude 2A (were A is a given number between 0 and π). Find the value of the electric field in the center of the arc. 2. Relevant equations $E(P)= \int K_e (dq/r^2)$ Where $K_e$ is the electrostatic constant 3. The attempt at a solution Since the charge is uniformly distributed I'm considering the charge density as λ= Q/L thus Q= λL, where L = 2A. The point P is the center of the arc. So the integral becomes $E(P)= \int K_e (λL/R^2)$ Getting out the constants, integrating on dL and selecting the limits 0 - L=2A $E(P)= K_e λ/R^2 \int_0^L LdL$ And solving: $E(P)= 2A K_e λ / R^2$ However, if I leave it in terms of Q and Pi, the expression oversimplifies and the term 2A goes away $E(P)= 2A Q / 4πε_0 R^2 2A$ $=Q/4πε_0 R^2$ So I got the serious feeling I'm getting something wrong here. Perhaps integrate in terms of the angle? 2. Jul 23, 2017 ### haruspex E is a vector. You need to integrate it as such. An integral is a sum. How do you add vectors? 3. Jul 23, 2017 ### SetepenSeth I think I'll start from scratch cause I don't follow 4. Jul 23, 2017 ### haruspex For example, consider the case where the arc goes through 180 degrees. The field at the centre from the charge at one end of the arc is equal and opposite to that from the charge at the other end of the arc, so those cancel. Choose a coordinate system and write the field as components in that system. You can then integrate each component separately.
2018-02-22 13:22:33
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9163107872009277, "perplexity": 593.5450447000799}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-09/segments/1518891814105.6/warc/CC-MAIN-20180222120939-20180222140939-00577.warc.gz"}
https://socratic.org/questions/how-do-you-write-10-3-1000-in-log-form
# How do you write 10^3 = 1000 in log form? ${\log}_{10} 1000 = 3$ or simply $\log 1000 = 3$ $y = {\log}_{a} x \iff {a}^{y} = x$
2020-04-07 17:44:22
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 3, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2525874972343445, "perplexity": 1384.2599072353355}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585371803248.90/warc/CC-MAIN-20200407152449-20200407182949-00545.warc.gz"}
https://no.overleaf.com/latex/templates/latex-template-for-the-dialogue-conference/trbnzvkptsgd
# LaTeX Template for the Dialogue Conference Author Dmitry Ustalov, Katya Artemova License Creative Commons CC BY 4.0 AbstractDialogue Conference paper in LaTeX filled in with some ideas on who to write a paper on headline generation
2020-10-20 06:30:17
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 1, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 1, "x-ck12": 0, "texerror": 0, "math_score": 0.24276874959468842, "perplexity": 10638.788746082231}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107869933.16/warc/CC-MAIN-20201020050920-20201020080920-00333.warc.gz"}
http://clay6.com/qa/18952/what-are-the-shapes-hybridization-of-so-2-so-3
Answer Comment Share Q) # What are the shapes & hybridization of $SO_2$ & $SO_3$ $\begin{array}{1 1}(a)\;SO_2-Angular,sp^2,SO_3-trigonal\;planar,sp^2&(b)\;\;SO_3-Angular,sp^2,SO_3-trigonal\;planar,sp^2\\(c)\;SO_2-linear\;sp,SO_3-trigonal\;planar,sp^2&(d)\;SO_2-SP^3,pyramid,SO_3-sp^2,linear\end{array}$
2020-02-26 14:04:42
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8053420186042786, "perplexity": 1908.0361095724872}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875146342.41/warc/CC-MAIN-20200226115522-20200226145522-00177.warc.gz"}
https://jeremykun.com/tag/capacitors/
# Taylor Series and Accelerometers In my book, A Programmer’s Introduction to Mathematics, I describe the Taylor Series as a “hammer for every nail.” I learned about another nail in the design of modern smartphone accelerometers from “Eight Amazing Engineering Stories” by Hammack, Ryan, and Ziech, which I’ll share here. These accelerometers are designed using a system involving three plates, which correspond to two capacitors. A quick recap on my (limited) understanding of how capacitors work. A capacitor involving two conductive plates looks like this: The voltage provided by the battery pushes electrons along the negative direction, or equivalently pushing “charge” along the positive direction (see the difference between charge flow and election flow). These elections build up in the plate labeled $-Q$, and the difference in charge across the two plates generates an electric field. If that electric field is strong enough, the electrons can jump the gap to the positive plate and complete the circuit. Otherwise, the plate reaches “capacity” and current stops flowing. Whether the jump happens or the current stops depends on the area of the plate $A$, the distance between the plates $d$, and the properties of the material between the plates, the last one is called the “dielectric constant” $\varepsilon$. (Nb., I’m not sure why it doesn’t depend on the material the plate is composed of, but I imagine it’s smooshed into the dielectric constant if necessary) This relationship is summarized by the formula $\displaystyle C = \frac{\varepsilon A}{d}$ Then, an external event can cause the plates to move close enough together so that the electrons can jump the gap and current can begin to flow. This discharges the negatively charged plate. A naive, Taylor-series-free accelerometer could work as follows: 1. Allow the negatively charged plate to wobble a little bit by fixing just one end of the plate, pictured like a diving board (a cantilever). 2. The amount of wobble will be proportional to the force of acceleration due to Hooke’s law for springs. 3. When displaced by a distance of $\delta$, the capacitance in the plate changes to $C = \frac{\varepsilon A}{d - \delta}$. 4. Use the amount of discharge to tell how much the plate displaced. This is able to measure the force of acceleration in one dimension, and so thee of these devices are arranged in perpendicular axes to allow one to measure acceleration in 3-dimensional space. The problem with this design is that $C = \frac{\varepsilon A}{d - \delta}$ is a nonlinear change in capacitance with respect to the amount of displacement. To see how nonlinear, expand this as a Taylor series: \displaystyle \begin{aligned} C &= \frac{\varepsilon A}{d - \delta} \\ &= \frac{\varepsilon A}{d} \left ( \frac{1}{1 - \frac{\delta}{d}} \right ) \\ &= \frac{\varepsilon A}{d} \left ( 1 + \frac{\delta}{d} + \left ( \frac{\delta}{d} \right )^2 + O_{\delta \to 0}(\delta^3) \right ) \end{aligned} I’m using the big-O notation $O_{\delta \to 0}(\delta^3)$ to more rigorously say that I’m “ignoring” all cubic and higher terms. I can do this because in these engineering systems (I’m taking Hammack at his word here), the quantity $(\delta / d)^2$ is meaningfully large, but later terms like $(\delta / d)^3$ are negligibly small. Of course, this is only true when the displacement $\delta$ is very small compared to $d$, which is why the big-O has a subscript $\delta \to 0$. Apparently, working backwards through the nonlinearity in the capacitance change is difficult enough to warrant changing the design of the system. (I don’t know why this is difficult, but I imagine it has to do with the engineering constraints of measurement devices; please do chime in if you know more) The system design that avoids this is a three-plate system instead of a two-plate system. In this system, the middle plate moves back and forth between two stationary plates that are connected to a voltage source. As it moves away from one and closer to the other, the increased capacitance on one side is balanced by the decreased capacitance on the other. The Taylor series shows how these two changes cancel out on the squared term only. If $C_1 = \frac{\varepsilon A}{d - \delta}$ represents the changed capacitance of the left plate (the middle plate moves closer to it), and $C_2 = \frac{\varepsilon A}{d + \delta}$ represents the right plate (the middle plate moves farther from it), then we expand the difference in capacitances via Taylor series (using the Taylor series for $1/(1-x)$ for both, but in the $1 + \delta/d$ case it’s $1 / (1 - (-x))$). \displaystyle \begin{aligned} C_1 - C_2 &= \frac{\varepsilon A}{d - \delta} - \frac{\varepsilon A}{d + \delta} \\ &= \frac{\varepsilon A}{d} \left ( \frac{1}{1 - \frac{\delta}{d}} - \frac{1}{1 + \frac{\delta}{d}} \right ) \\ &= \frac{\varepsilon A}{d} \left ( 1 + \frac{\delta}{d} + \left ( \frac{\delta}{d} \right )^2 + O_{\delta \to 0}(\delta^3) - 1 + \frac{\delta}{d} - \left ( \frac{\delta}{d} \right )^2 + O_{\delta \to 0}(\delta^3) \right ) \\ &= \frac{\varepsilon A}{d} \left ( \frac{2\delta}{d} + O_{\delta \to 0}(\delta^3) \right ) \end{aligned} Again, since the cubic and higher terms are negligibly small, we can “ignore” those parts. What remains is a linear response to the change in the middle plate’s displacement. This makes it significantly easier to measure. Because we’re measuring the difference in capacitance, this design is called a “differential capacitor.” Though the math is tidy in retrospect, I marvel at how one might have conceived of this design from scratch. Did the inventor notice the symmetries in the Taylor series approximations could be arranged to negate each other? Was there some other sort of “physical intuition” at play? Until next time!
2020-08-10 02:30:23
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 21, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.99439537525177, "perplexity": 594.5029054243913}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439738603.37/warc/CC-MAIN-20200810012015-20200810042015-00423.warc.gz"}
https://www.shaalaa.com/question-bank-solutions/pythagoras-theorem-diagonals-rectangle-equal-bisect-each-othe_1872
# Solution - Pythagoras Theorem Account Register Share Books Shortlist ConceptPythagoras Theorem #### Question Prove that the diagonals of a rectangle ABCD, with vertices A(2, -1), B(5, -1), C(5, 6) and D(2, 6), are equal and bisect each other. #### Solution You need to to view the solution Is there an error in this question or solution? #### Similar questions VIEW ALL Sides of triangles are given below. Determine which of them are right triangles? In case of a right triangle, write the length of its hypotenuse. 3 cm, 8 cm, 6 cm view solution Prove that the points (3, 0), (6, 4) and (-1, 3) are the vertices of a right angled isosceles triangle. view solution Tick the correct answer and justify: In ΔABC, AB = 6sqrt3 cm, AC = 12 cm and BC = 6 cm. The angle B is: (A) 120° (B) 60° (C) 90° (D) 45° view solution ABC is an isosceles triangle right angled at C. Prove that AB2 = 2AC2 view solution If ABC is an equilateral triangle of side a, prove that its altitude =  \frac { \sqrt { 3 } }{ 2 } a view solution #### Reference Material Solution for concept: Pythagoras Theorem. For the course 8th-10th CBSE S
2017-10-23 15:20:06
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6967706680297852, "perplexity": 1630.5006021990953}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187826114.69/warc/CC-MAIN-20171023145244-20171023165244-00865.warc.gz"}
https://newproxylists.com/probability-whats-more-likely-7-digit-number-with-no-1s-or-at-least-one-1-among-its-digits/
# probability – What’s more likely: \$7\$-digit number with no \$1\$’s or at least one \$1\$ among its digits? A $$7$$-digit number is chosen at random. Which is more likely: the number has no $$1$$‘s among its digits or the number has at least one $$1$$ among its digits? Here’s how I did it: The question is asking whether $$8(9)^6$$ (the number of those with no $$1$$‘s among its digits) or $$9(10)^6 – 8(9)^6$$ (the number of those with at least one $$1$$ among its digits). Some tedious multiplying shows that $$8(9)^6 = 4241528 < 4500000$$, which demonstrates that $$9(10^6) – 8(9)^6$$ i.e. the number having at least one $$1$$‘s among its digits is more likely. However, I am wondering if there is a slicker way to get the answer without having to do any tedious multipication.
2021-07-24 20:52:01
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 10, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.843571126461029, "perplexity": 226.97453450960654}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046150308.48/warc/CC-MAIN-20210724191957-20210724221957-00216.warc.gz"}
http://www.oalib.com/relative/3405281
Home OALib Journal OALib PrePrints Submit Ranking News My Lib FAQ About Us Follow Us+ Title Keywords Abstract Author All Search Results: 1 - 10 of 100 matches for " " Page 1 /100 Display every page 5 10 20 Item Physics , 2014, DOI: 10.1016/j.elspec.2014.01.005 Abstract: We characterize the topological insulator Bi$_2$Se$_3$ using time- and angle- resolved photoemission spectroscopy. By employing two-photon photoemission, a complete picture of the unoccupied electronic structure from the Fermi level up to the vacuum level is obtained. We demonstrate that the unoccupied states host a second, Dirac surface state which can be resonantly excited by 1.5 eV photons. We then study the ultrafast relaxation processes following optical excitation. We find that they culminate in a persistent non-equilibrium population of the first Dirac surface state, which is maintained by a meta-stable population of the bulk conduction band. Finally, we perform a temperature-dependent study of the electron-phonon scattering processes in the conduction band, and find the unexpected result that their rates decrease with increasing sample temperature. We develop a model of phonon emission and absorption from a population of electrons, and show that this counter-intuitive trend is the natural consequence of fundamental electron-phonon scattering processes. This analysis serves as an important reminder that the decay rates extracted by time-resolved photoemission are not in general equal to single electron scattering rates, but include contributions from filling and emptying processes from a continuum of states. Physics , 2012, DOI: 10.1103/PhysRevLett.108.185501 Abstract: In this letter we report measurements of the coupling between Dirac fermion quasiparticles (DFQs) and phonons on the (001) surface of the strong topological insulator Bi2Se3. While most contemporary investigations of this coupling have involved examining the temperature dependence of the DFQ self-energy via angle-resolved photoemission spectroscopy (ARPES) measurements, we employ inelastic helium atom scattering to explore, for the first time, this coupling from the phonon perspective. Using a Hilbert transform, we are able to obtain the imaginary part of the phonon self-energy associated with a dispersive surface phonon branch identified in our previous work [1] as having strong interactions with the DFQs. From this imaginary part of the self-energy we obtain a branch-specific electron-phonon coupling constant of 0.43, which is stronger than the values reported form the ARPES measurements. Physics , 2004, Abstract: Recently there has been an accumulation of experimental evidence in the high temperature superconductors suggesting the relevance of electron-phonon coupling in these materials. These findings challenge some well-held beliefs of what electron-phonon interactions can and cannot do. In this article we review evidence primarily from angle-resolved photoemission (ARPES) measurements which point out the importance of electronic coupling to certain phonon modes in the cuprates. Physics , 2012, DOI: 10.1088/0268-1242/27/12/124001 Abstract: Angle-resolved photoelectron spectroscopy is used for a detailed study of the electronic structure of the topological insulator Bi2Se3. Nominally stoichiometric and calcium-doped samples were investigated. The pristine surface shows the topological surface state in the bulk band gap. As time passes, the Dirac point moves to higher binding energies, indicating an increasingly strong downward bending of the bands near the surface. This time-dependent band bending is related to a contamination of the surface and can be accelerated by intentionally exposing the surface to carbon monoxide and other species. For a sufficiently strong band bending, additional states appear at the Fermi level. These are interpreted as quantised conduction band states. For large band bendings, these states are found to undergo a strong Rashba splitting. The formation of quantum well states is also observed for the valence band states. Different interpretations of similar data are also discussed. Physics , 2006, Abstract: This treatise reviews latest results obtained from angle-resolved photoemission spectroscopy (ARPES) on cuprate superconductors, with a special focus on the electron-phonon interaction. What has emerged is rich information about the anomalous electron-phonon interaction well beyond the traditional views of the subject. It exhibits strong doping, momentum and phonon symmetry dependence, and shows complex interplay with the strong electron-electron interaction in these materials. Physics , 2013, DOI: 10.1088/1367-2630/15/8/083023 Abstract: Based on results from femtosecond time-resolved photoemission, we compare three different methods for determination of the electron-phonon coupling constant {\lambda} in Eu and Ba-based 122 FeAs compounds. We find good agreement between all three methods, which reveal a small {\lambda} < 0.2. This makes simple electron-phonon mediated superconductivity unlikely in these compounds. Physics , 2011, DOI: 10.1103/PhysRevLett.108.187001 Abstract: Gapless surface states on topological insulators are protected from elastic scattering on non-magnetic impurities which makes them promising candidates for low-power electronic applications. However, for wide-spread applications, these states should have to remain coherent at ambient temperatures. Here, we studied temperature dependence of the electronic structure and the scattering rates on the surface of a model topological insulator, Bi$_2$Se$_3$, by high resolution angle-resolved photoemission spectroscopy. We found an extremely weak broadening of the topological surface state with temperature and no anomalies in the state's dispersion, indicating exceptionally weak electron-phonon coupling. Our results demonstrate that the topological surface state is protected not only from elastic scattering on impurities, but also from scattering on low-energy phonons, suggesting that topological insulators could serve as a basis for room temperature electronic devices. Physics , 2009, DOI: 10.1103/PhysRevB.81.041403 Abstract: The electron-phonon coupling in potassium-doped graphene on Ir(111) is studied via the renormalization of the pi* band near the Fermi level, using angle-resolved photoemission spectroscopy. The renormalization is found to be fairly weak and almost isotropic, with a mass enhancement parameter of lambda= 0.28(6) for both the K-M and the K-G direction. These results are found to agree well with recent first principles calculations. Physics , 2007, Abstract: We have studied the O 2p valence-band structure of Nb-doped SrTiO3, in which a dilute concentration of electrons are doped into the d0 band insulator, by angle-resolved photoemission spectroscopy (ARPES) measurements. We found that ARPES spectra at the valence band maxima at the M [k = (pi/a, pi/a, 0)]and R [k = (pi/a, pi/a, pi/a)] points start from ~ 3.3 eV below the Fermi level (EF), consistent with the indirect band gap of 3.3 eV and the EF position at the bottom of the conduction band. The peak position of the ARPES spectra were, however, shifted toward higher binding energies by ~ 500 meV from the 3.3 eV threshold. Because the bands at M and R have pure O 2p character, we attribute this ~ 500 meV shift to strong coupling of the oxygen p hole with optical phonons in analogy with the peak shifts observed for d-electron photoemission spectra in various transition-metal oxides. Physics , 2007, DOI: 10.1103/PhysRevB.75.195116 Abstract: Lattice contribution to the electronic self-energy in complex correlated oxides is a fascinating subject that has lately stimulated lively discussions. Expectations of electron-phonon self-energy effects for simpler materials, such as Pd and Al, have resulted in several misconceptions in strongly correlated oxides. Here we analyze a number of arguments claiming that phonons cannot be the origin of certain self-energy effects seen in high-$T_c$ cuprate superconductors via angle resolved photoemission experiments (ARPES), including the temperature dependence, doping dependence of the renormalization effects, the inter-band scattering in the bilayer systems, and impurity substitution. We show that in light of experimental evidences and detailed simulations, these arguments are not well founded. Page 1 /100 Display every page 5 10 20 Item
2020-01-18 12:51:10
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6726193428039551, "perplexity": 1661.8216355485335}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-05/segments/1579250592565.2/warc/CC-MAIN-20200118110141-20200118134141-00003.warc.gz"}
http://nkbrasil.com.br/fidelity-investments-aabxo/21cdc5-discrete-function-calculator
$P(X=x)=k$ for $x=4,5,6,7,8$, where $k$ is constant. inverse f ( x) = cos ( 2x + 5) $inverse\:f\left (x\right)=\sin\left (3x\right)$. The input to the floor function is any real number x and its output is the greatest integer less than or equal to x. \end{aligned} $$, Now, Variance of discrete uniform distribution X is,$$ \begin{aligned} V(X) &= E(X^2)-[E(X)]^2\\ &=100.67-[10]^2\\ &=100.67-100\\ &=0.67. Compute answers using Wolfram's breakthrough technology & knowledgebase, relied on by millions of students & professionals. To analyze our traffic, we use basic Google Analytics implementation with anonymized data. The probability mass function (pmf) of $X$ is, \begin{aligned} P(X=x) &=\frac{1}{5-0+1} \\ &= \frac{1}{6}; x=0,1,2,3,4,5. Given a discrete random variable, $$X$$, its probability distribution function, $$f(x)$$, is a function that allows us to calculate the probability that $$X=x$$. The probability mass function (pmf) of X is, \begin{aligned} P(X=x) &=\frac{1}{11-9+1} \\ &= \frac{1}{3}; x=9,10,11. If the function is one-to-one, there will be a unique inverse. f ( x) = x3. Step 4 - Click on “Calculate” for discrete uniform distribution, Step 6 - Calculate cumulative probabilities. Find more Mathematics widgets in Wolfram|Alpha. Convolution calculator online. 1. Free online calculators for exponents, math, fractions, factoring, plane geometry, solid geometry, algebra, finance and trigonometry Calculator,Discrete Uniform distribution, Discrete uniform distribution examples, Discrete uniform distribution calculator, uniform distribution definition,mean,variance If the function is one-to-one, there will be a unique inverse. \end{aligned} $$, Mean of discrete uniform distribution X is,$$ \begin{aligned} E(X) &=\sum_{x=9}^{11}x \times P(X=x)\\ &= \sum_{x=9}^{11}x \times\frac{1}{3}\\ &=9\times \frac{1}{3}+10\times \frac{1}{3}+11\times \frac{1}{3}\\ &= \frac{9+10+11}{3}\\ &=\frac{30}{3}\\ &=10. Step 1: Enter the expression you want to evaluate. To configure the filter for discrete time, set the Sample time property to a positive, nonzero value, or to -1 to inherit the sample time from an upstream block. y = x2 + x + 1 x. \end{aligned} $$, And variance of discrete uniform distribution Y is,$$ \begin{aligned} V(Y) &=V(20X)\\ &=20^2\times V(X)\\ &=20^2 \times 2.92\\ &=1168. \end{aligned} . 3. It is a tool used to convert the finite sequence of equally-spaced samples of any function into an equivalent-length sequence. Comparing the efficiently of different algorithms that solve the same problem. There is an extremely powerful tool in discrete mathematics used to manipulate sequences called the generating function. \end{aligned}, \begin{aligned} E(X) &=\sum_{x=0}^{5}x \times P(X=x)\\ &= \sum_{x=0}^{5}x \times\frac{1}{6}\\ &=\frac{1}{6}(0+1+2+3+4+5)\\ &=\frac{15}{6}\\ &=2.5. The identity function f(z)=z in the complex plane is illustrated above. \end{aligned}, The variance of discrete uniform distribution $X$ is, \begin{aligned} V(X) &=\frac{(8-4+1)^2-1}{12}\\ &=\frac{25-1}{12}\\ &= 2 \end{aligned}, c. The probability that $X$ is less than or equal to 6 is, \begin{aligned} P(X \leq 6) &=P(X=4) + P(X=5) + P(X=6)\\ &=\frac{1}{5}+\frac{1}{5}+\frac{1}{5}\\ &= \frac{3}{5}\\ &= 0.6 \end{aligned}. c. The mean of discrete uniform distribution $X$ is, \begin{aligned} E(X) &=\frac{1+6}{2}\\ &=\frac{7}{2}\\ &= 3.5 \end{aligned} Hope you like article on Discrete Uniform Distribution. Derivative numerical and analytical calculator A4:A11 in Figure 1) and R2 is the range consisting of the frequency values f(x) corresponding to the x values in R1 (e.g. Below are the few solved examples on Discrete Uniform Distribution with step by step guide on how to find probability and mean or variance of discrete uniform distribution. By millions of students & professionals equivalent-length sequence → 4: Status → about ID may look like: BD92F455... ( pmf ) of discrete mathematics calculators a uniform amount ( annuity -! Most important discrete transform used to perform Fourier analysis in various practical applications top of a die changing. To the discrete function calculator function: transfer functions are a frequency-domain representation of time-invariant! Basically, it shows how many different possible subsets can be made from the larger set blog, Wordpress Blogger... Remainder from a larger set a uniform amount ( annuity ) - … Section 5.1 functions.: Enter the expression you want to evaluate a periodic function and thus can not any... Used to manipulate sequences called the standard deviation mass function ( pmf ) of discrete uniform distribution Step! 6X + 8 } { x^2 } $if needed an absolute value equation using the following steps: the... A tool used to convert the finite sequence of complex numbers into … Enter the last digit of selected! For variance, and standard deviation for a given value of discrete variable... All cookies on the integers$ 0\leq x\leq 5 $of the and. Continuous uniform distribution calculator, the order of the items chosen in the subset does not matter half a. It shows how many different possible subsets can be obtained by taking sample...: Status → about ID may look like: 1008000007206E210B0 BD92F455 - Click on “Calculate” for discrete uniform distribution to! To manually perform the discrete representation is equivalent to 5 * x that you are to... You continue without changing your Settings, we need to calculate$ E ( )! 5X is equivalent to 5 * x remainder from a specific.. For the most important discrete Distributions in mathematics range defining the discrete Fourier (... + 3 is illustrated above ( DFT ) $are equally likely$, let Y=20X! A continuous graph follows a discrete uniform distribution, Step 6 - calculate probabilities! The integers $9\leq x\leq 11$ are $0,1,2, \cdots, 9$ are equally likely evaluate problem. $9, 10, 11$ are equally likely sign, so ., combinations, replacements, nCr and nPr calculators Distributions which contains calculators factorials! That $x\leq 6$ 0,9 ) $standard deviation equally likely problem down a. We need to calculate$ E ( x^2 ) $inverse\: f\left ( x\right ) =\sqrt { x+3$. Function f ( x − 5 ) $will be a unique inverse 9, 10 11. √X + 3 the absolve value expression by itself or expected value of discrete random variable is defined as:... 8 digits of your 27-digit TI-Nspire 's Product ID: it is the place for you as you will all! X ( e.g factorials, odd and even permutations, combinations discrete function calculator replacements, nCr and nPr.... Let the random variable$ x $have a discrete uniform distribution on the number two given a remainder a..., let$ x $denote the number two given a remainder from a set. Students & professionals last 8 digits of your 27-digit TI-Nspire 's Product ID cos ( 2x + 5$. Logarithms based on the integers $0\leq x\leq 5$ is an extremely powerful tool in discrete mathematics used perform., \cdots, 9 $are equally likely calculate basic discrete random variable$ x $is one-to-one, will. Square root of the functions, with steps shown to provide a comment feature ( x-5\right )$ you skip... Receive all cookies on the top is less than or equal to x 5: →... ( annuity ) - … Section 5.1 Generating functions discrete transform used to manipulate sequences called the standard.. All cookies on the integers $9\leq x\leq 11$ there will be a discrete uniform $... The possible values of the people remaining on the integers$ 0\leq x\leq 5 $that be... Composition at the specified discrete function calculator, if needed function into an equivalent-length sequence =z in subset. Discrete representation is equivalent to 5 * x a unique inverse discrete function calculator, 11$ $. 2X+5\Right )$ technology & knowledgebase, relied on by millions of students professionals... Into an equivalent-length sequence $x$ denote the number of possible combinations that can be made from larger. ) - … Section 5.1 Generating functions so you can better understand the results provided by this calculator can you! An online calculator to find probability and cumulative probabilities want to evaluate 's breakthrough technology knowledgebase! The Math calculator will find the composition of the floor function is any real number and. You get the free discrete function '' widget for your website, blog, Wordpress Blogger! Evaluate your problem down to a final solution given value of discrete uniform distribution on the top a! Perform the discrete values of the floor and ceiling functions for a given value of discrete random $! Appear on the top of a die PROB, which is defined follows...: Status → about ID may look like: 1008000007206E210B0 BD92F455 order the... X mass function ( pmf ) of discrete mathematics used to convert the finite sequence of numbers! Same problem = sin ( 3x ) function-inverse-calculator follows a discrete graph, not a continuous graph Uses! Prob, which is the probability that$ x\leq 6 $11$ and its is. By itself $0,1,2,3,4,5$ are $0,1,2, \cdots, 9$ of randomly selected telephone number is tool... Policy | Terms of Use is an extremely powerful tool in discrete mathematics calculators 6x + 8 z ) in! Distribution $U ( 0,9 )$: Settings → 4: Status → about ID may look:... Last digit of randomly selected telephone number discrete Fourier Series ( DFT ) - Click on “Calculate” discrete! An absolute value equation using the following steps: get the absolve value expression itself... Are a frequency-domain representation of linear time-invariant systems odd and even permutations, combinations, replacements, nCr nPr. Mass function ( pmf ) of discrete mathematics calculators as you will get kinds. To find probability and cumulative discrete function calculator ) $y=\frac { x } { }. Of selected telephone number for a probability distribution so you can better understand the provided. 3X\Right )$ f\left ( x\right ) =\cos\left ( 2x+5\right ) E ( x^2 ) $discrete... Your 27-digit TI-Nspire 's Product ID x taking a sample of items from a set. Periodic function and thus can not represent any arbitrary function x x2 − 6x + 8 most discrete! ) function-inverse-calculator is one-to-one, there will be a discrete uniform distribution calculator to probability. You continue without changing your Settings, we need to calculate basic discrete random variable metrics mean! A continuous graph variable is defined as down to a final solution discrete random variable x... Manually perform the discrete representation is equivalent to 5 * x 5: →... Distribution, Step 6 - calculate cumulative probabilities Use the DFT Formula to manually the! The equation Click on “Calculate” for discrete uniform distribution on the island would be a discrete uniform calculator... 9$ you will get all kinds of discrete discrete function calculator calculators get all kinds of discrete random is. Y=20X $all cookies on the island would be a unique inverse practical applications hopefully, half of die! =\Cos\Left ( 2x+5\right )$ by millions of students & professionals absolve value expression by itself, combinations replacements! Settings → 4: Status → discrete function calculator ID may look like: 1008000007206E210B0 BD92F455 Us | our Team Privacy. C. compute mean and standard deviation for a probability distribution so you can better understand the provided... Under 5: Settings → 4: Status → about ID may look like 1008000007206E210B0! To find probability and cumulative probabilities a frequency-domain representation of linear time-invariant systems a die equation using the steps! Value of the random variable is defined as follows: a continuous graph and its is. Any function into an equivalent-length sequence to manipulate sequences called the standard deviation follows.. To manually perform the discrete Fourier Series ( DFT ) called the Generating function of selected telephone.. The following steps: get the absolve value expression by itself if needed the vrcacademy.com website Analytics with! Practical applications is less than 3. c. compute mean and variance of $x$ vrcacademy.com.! Dft Uses: it is a tool used to convert the finite sequence of discrete function calculator numbers into … the. This website Uses cookies to ensure you get the free ` discrete function '' widget for website! Variable x ( e.g any real number x and its output is probability... Can better understand the results provided by this calculator, the order of the items chosen in the plane... Discrete probability Distributions which contains calculators for the most important discrete transform used to manipulate called. And analytical calculator calculator ; Formula ; Use the DFT Formula to manually the... - Step by Step Guide, Karl Pearson coefficient of skewness for grouped.... Cookies to ensure you get the best experience on our site and to provide a comment feature number! Place for you as you will get all kinds of discrete uniform distribution + 8 combinations calculator evaluate! And to provide a comment feature a comment feature this is the most important discrete used. Analytical calculator discrete function calculator ; Formula ; Use the DFT Formula to manually perform discrete!: f\left ( x\right ) =\ln\left ( x-5\right ) \$ f\left ( x\right ) =\sin\left ( ). Taking a sample of items from a larger set calculator calculator ; Formula ; the... Y = x x2 − 6x + 8 Us | our Team | Policy. As follows: calculator calculator ; Formula ; Use the DFT Formula to perform! Maria Ross Red Slice, Graphic Design Prompt Generator, American College Of Dubai Fees, His Administration Believed That Was Vital To The Country, Losing Weight But Arms Still Fat, Brookfield Asset Management Case Study, Swivel Car Seats For Sale, South African Palm Fruit, Leading Causes Of Death Vegan, Renault Fluence Ze Battery Replacement, From The Ground Up Cauliflower Stars, Family Mart Outlet, Distributive Property With Variables Calculator, Desktop Backgrounds 1920x1080,
2021-09-27 06:17:49
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 2, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9240285754203796, "perplexity": 1762.2326053332074}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780058373.45/warc/CC-MAIN-20210927060117-20210927090117-00027.warc.gz"}
https://www.esaral.com/q/solve-this-following-34337
Deepak Scored 45->99%ile with Bounce Back Crack Course. You can do it too! # Solve this following Question: If $(2 A-B)=\left(\begin{array}{ccc}6 & -6 & 0 \\ -4 & 2 & 1\end{array}\right)$ and $(2 B+A)=\left(\begin{array}{lll}3 & 2 & 5 \\ -2 & 1 & -7\end{array}\right)$ then $A=?$ A. $\left(\begin{array}{llr}-3 & 2 & 1 \\ 2 & 1 & -1\end{array}\right)$ B. $\left(\begin{array}{ccc}3 & 2 & -1 \\ 2 & -1 & 1\end{array}\right)$ C. $\left(\begin{array}{lll}3 & -2 & 1 \\ -2 & 1 & -1\end{array}\right)$ D. none of these Solution:
2023-01-28 01:34:35
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6815366148948669, "perplexity": 4679.562939755225}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499468.22/warc/CC-MAIN-20230127231443-20230128021443-00642.warc.gz"}
https://search.r-project.org/CRAN/refmans/dtangle/html/process_markers.html
process_markers {dtangle} R Documentation ## Determines number of markers n_markers, marker list mrkrs, and gamma. ### Description Determines number of markers n_markers, marker list mrkrs, and gamma. ### Usage process_markers(Y, pure_samples, n_markers, data_type, gamma, markers, marker_method) ### Arguments Y Expression matrix. (Required) Two-dimensional numeric. Must implement as.matrix. Each row contains expression measurements for a particular sample. Each columm contains the measurements of the same gene over all individuals. Can either contain just the mixture samples to be deconvolved or both the mixture samples and the reference samples. See pure_samples and references for more details. pure_samples The pure sample indicies. (Optional) List of one-dimensional integer. Must implement as.list. The i-th element of the top-level list is a vector of indicies (rows of Y or references) that are pure samples of type i. If references is not specified then this argument identifies which rows of Y correspond to pure reference samples of which cell-types. If references is specified then this makes same idenficiation but for the references matrix instead. n_markers Number of marker genes. (Optional) One-dimensional numeric. How many markers genes to use for deconvolution. Can either be a single integer, vector of integers (one for each cell type), or single or vector of percentages (numeric in 0 to 1). If a single integer then all cell types use that number of markers. If a vector then the i-th element determines how many marker genes are used for the i-th cell type. If single percentage (in 0 to 1) then that percentage of markers are used for all types. If vector of percentages then that percentage used for each type, respectively. If not specified then top 10% of genes are used. data_type Type of expression measurements. (Optional) One-dimensional string. An optional string indicating the type of the expression measurements. This is used to set gamma to a pre-determined value based upon the data type. Valid values are for probe-level microarray as “microarray-probe”, gene-level microarray as “microarray-gene” or rna-seq as “rna-seq”. Alternatively can set gamma directly. gamma Expression adjustment term. (Optional) One-dimensional positive numeric. If provided as a single positive number then that value will be used for gamma and over-ride the value of gamma chosen by the data_type argument. If neither gamma nor data_type are specified then gamma will be set to one. markers Marker gene indices. (Optional) List of one-dimensional integer. Top-level list should be same length as pure_samples, i.e. one element for each cell type. Each element of the top-level list is a vector of indicies (columns of Y) that will be considered markers of that particular type. If not supplied then dtangle finds markers internally using find_markers. Alternatively, one can supply the output of find_markers to the markers argument. marker_method Method used to rank marker genes. (Optional) One-dimensional string. The method used to rank genes as markers. If not supplied defaults to “ratio”. Only used if markers are not provided to argument “markers”. Options are 'ratio' selects and ranks markers by the ratio of the mean expression of each gene in each cell type to the mean of that gene in all other cell types. 'regression ' selects and ranks markers by estimated regression coefficients in a series of regressions with single covariate that is indicator of each type. 'diff' selects and ranks markers based upon the difference, for each cell type, between the median expression of a gene by each cell type and the median expression of that gene by the second most highly expressed cell type. 'p.value' selects and ranks markers based upon the p-value of a t-test between the median expression of a gene by each cell type and the median expression of that gene by the second most highly expressed cell type. [Package dtangle version 2.0.9 Index]
2022-05-16 18:18:13
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7191974520683289, "perplexity": 3537.866517904532}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652662512229.26/warc/CC-MAIN-20220516172745-20220516202745-00707.warc.gz"}
http://mathhelpforum.com/advanced-algebra/27222-find-linear-transformation-print.html
# Find a linear transformation • Feb 1st 2008, 07:37 AM Find a linear transformation Is there a linear transformation $T:R^3 \rightarrow R^2$ such that T(1,0,3) = (1,1) and T(-2,0,-6) = (2,1)? I set the system of 1x+3y = 1 and -2x + -6y = 2 in an attempt to find x and y, but it turns out to be no solution, does that means there is no linear transformation in this case? • Feb 1st 2008, 07:40 AM topsquark Quote: Is there a linear transformation $T:R^3 \rightarrow R^2$ such that T(1,0,3) = (1,1) and T(-2,0,-6) = (2,1)? I set the system of 1x+3y = 1 and -2x + -6y = 2 in an attempt to find x and y, but it turns out to be no solution, does that means there is no linear transformation in this case? Sounds about right to me. :) -Dan • Feb 1st 2008, 10:34 AM (-2)(1,0,3) = (-2,0,-6) but then (-2)T(x) $\neq$ T(-2x).
2016-09-28 07:50:14
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 3, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6491426229476929, "perplexity": 842.5651263418843}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-40/segments/1474738661326.74/warc/CC-MAIN-20160924173741-00039-ip-10-143-35-109.ec2.internal.warc.gz"}
http://gmatclub.com/forum/paul-has-twenty-five-transit-cards-each-worth-either-112129.html
Find all School-related info fast with the new School-Specific MBA Forum It is currently 30 May 2016, 11:23 ### GMAT Club Daily Prep #### Thank you for using the timer - this advanced tool can estimate your performance and suggest more practice questions. We have subscribed you to Daily Prep Questions via email. Customized for You we will pick new questions that match your level based on your Timer History Track every week, we’ll send you an estimated GMAT score based on your performance Practice Pays we will pick new questions that match your level based on your Timer History # Events & Promotions ###### Events & Promotions in June Open Detailed Calendar # Paul has twenty-five transit cards, each worth either.. Author Message TAGS: ### Hide Tags Manager Joined: 16 Mar 2011 Posts: 177 Followers: 1 Kudos [?]: 35 [0], given: 13 Paul has twenty-five transit cards, each worth either.. [#permalink] ### Show Tags 10 Apr 2011, 03:55 00:00 Difficulty: (N/A) Question Stats: 0% (00:00) correct 0% (00:00) wrong based on 0 sessions ### HideShow timer Statistics Paul has twenty-five transit cards, each worth either$5,$3, or \$1.50. What is the total monetary value of all of Paul's transit cards? values of cards: 5x + 3y + 1.50(25 - x - y) = ? Simplify ~ 3.5x + 1.5y + 37.5 = ? From here I tried solving for x and y.. Here's what I did, please do correct me if I'm wrong (most probably am ) y = (-37,5 - 3,5x)/1,5 so, 3,5x + 1,5(-37,5 - 3,5x)/1,5 + 37,5 = 3,5x + -37,5 - 3,5x + 37,5 = x = 0 And then I got stuck SVP Joined: 16 Nov 2010 Posts: 1673 Location: United States (IN) Concentration: Strategy, Technology Followers: 34 Kudos [?]: 442 [0], given: 36 Re: Paul has twenty-five transit cards, each worth either.. [#permalink] ### Show Tags 10 Apr 2011, 04:19 Hi I think we need more information. This example is present there just to illustrate how to frame/imbibe maximum information from question even before reading the DS choices. "All of the work so far has come just from the prompt of a Data Sufficiency question—you have not even seen statements (1) and (2) yet! But this work is worth the time and energy. In general, you should rephrase and interpret a Data Sufficiency question prompt as much as you can before you begin to work with the statements." Btw, I have not gone beyond this text to read any other information, so please feel free to dig further. Regards, Subhash _________________ Formula of Life -> Achievement/Potential = k * Happiness (where k is a constant) GMAT Club Premium Membership - big benefits and savings Manager Joined: 16 Mar 2011 Posts: 177 Followers: 1 Kudos [?]: 35 [0], given: 13 Re: Paul has twenty-five transit cards, each worth either.. [#permalink] ### Show Tags 10 Apr 2011, 04:21 Hi, yeah so at first I thought that I can't solve it because I got 2 variables and 1 equation but then my friend (who is a math genius) said that it could be solved so I tried doing so and ended up being more confused.. SVP Joined: 16 Nov 2010 Posts: 1673 Location: United States (IN) Concentration: Strategy, Technology Followers: 34 Kudos [?]: 442 [0], given: 36 Re: Paul has twenty-five transit cards, each worth either.. [#permalink] ### Show Tags 10 Apr 2011, 04:23 Oh yeah, let us wait to hear from Gurus as Walker, Ian, fluke etc. _________________ Formula of Life -> Achievement/Potential = k * Happiness (where k is a constant) GMAT Club Premium Membership - big benefits and savings Re: Paul has twenty-five transit cards, each worth either..   [#permalink] 10 Apr 2011, 04:23 Similar topics Replies Last post Similar Topics: 13 A certain game pays players in tokens, each of which is worth either m 7 11 May 2015, 06:07 3 In blackjack, each card is worth a certain amount of points. 3 30 Apr 2013, 21:37 20 In a certain deck of cards, each card has a positive integer 9 11 Apr 2012, 02:34 3 In a certain deck of cards, each card has a positive integer written 7 24 Aug 2011, 07:28 1 A deck of 52 cards is dealt out to 4 players (each player ha 4 10 Aug 2010, 10:33 Display posts from previous: Sort by
2016-05-30 18:23:51
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2623983919620514, "perplexity": 5336.801345038218}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-22/segments/1464051054181.34/warc/CC-MAIN-20160524005054-00209-ip-10-185-217-139.ec2.internal.warc.gz"}
https://jgrapht.org/javadoc-1.3.0/org/jgrapht/alg/interfaces/AStarAdmissibleHeuristic.html
org.jgrapht.alg.interfaces • Type Parameters: V - vertex type All Known Implementing Classes: public interface AStarAdmissibleHeuristic<V> Interface for an admissible heuristic used in A* search. Author: Joris Kinable, Jon Robison, Thomas Breitbart • ### Method Summary All Methods Modifier and Type Method and Description double getCostEstimate(V sourceVertex, V targetVertex) An admissible "heuristic estimate" of the distance from $x$, the sourceVertex, to the goal (usually denoted $h(x)$). • ### Method Detail • #### getCostEstimate double getCostEstimate(V sourceVertex, V targetVertex) An admissible "heuristic estimate" of the distance from $x$, the sourceVertex, to the goal (usually denoted $h(x)$). This is the good guess function which must never overestimate the distance. Parameters: sourceVertex - the source vertex targetVertex - the target vertex Returns: an estimate of the distance from the source to the target vertex
2023-03-29 23:01:43
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7357951998710632, "perplexity": 7307.935069060277}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296949035.66/warc/CC-MAIN-20230329213541-20230330003541-00002.warc.gz"}
https://shotaochi.github.io/release/2018/11/25/Adaptive_Double_Plateaus_Histogram_Equalization.html
Histogram equalization (HE) is a widely used technique for contrast enhancement of images. However, HE sometimes causes over-enhancement of background noise, whereas the contrast of objects is not enhanced well. Double plateaus histogram equalization (DPHE) was proposed for overcoming this problem. DPHE modifies original histogram as shown below, then equalizes the modified histogram. $H_{m}(g) = \begin{cases} T_{UP} & h(g) >= T_{UP} \\ h(g) & T_{DOWN} < h(g) < T_{UP} \\ T_{DOWN} & h(g) <= T_{DOWN} \end{cases}$ where $$g$$ is the gray level, $$h(g)$$ is the original histogram, and $$H_{m}(g)$$ is the modified histogram. The paramters, $$T_{DOWN}$$ and $$T_{UP}$$, are key parameters and its optimal values are scene specific. Tuning the parameters for each scene are annoying. That’s why adaptive double plateaus histogram equalization (ADPHE), which computes the parameters automatically, was proposed. devtools::install_github("ShotaOchi/imagerExtra") hist(ADPHE, main = "Histogram of ADPHE")
2023-03-25 00:24:52
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5341916084289551, "perplexity": 4816.841544140333}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296945292.83/warc/CC-MAIN-20230325002113-20230325032113-00508.warc.gz"}
https://www.ibm.com/blogs/research/2015/05/silicon-photonics-the-future-of-high-speed-data/
# Silicon Photonics: The future of High-Speed Data IBM’s research in brain-inspired computing, quantum computing, and silicon photonics is preparing to take computing in entirely new directions. The neuromorphic chip is getting smarter, the quantum bits are being scaled out, and in the near future, my team’s CMOS Integrated Nano-Photonics Technology will help ease data traffic jams in all sorts of computing and communications systems – pushing cloud computing and Big Data analytics to achieve their full potential. For the first time, we have designed and tested a fully integrated, wavelength-multiplexed silicon photonics chip capable of optically transmitting and receiving information at data rates up to 25 Gb/s per channel. This will soon make it possible to manufacture optical transceivers capable of transmitting 100 gigabits of data per second. Silicon photonics technology gives computational systems the ability to use pulses of light to move data at high speeds over optical fibers, instead of using conventional electrical signals over copper wires. Optical interconnects, based on vertical-cavity surface-emitting laser (VCSEL) technology and multi-mode fiber, are already being used in systems today. But their transmission range is limited to a relatively short distance of about 150 meters. Today, large data centers continue to scale in size to support exponentially growing traffic from social media, video streaming, cloud storage, sensor data, and much more. The longest optical links in such systems can be more than a kilometer in length. As a result, new optical interconnect solutions that can meet these requirements at low cost are needed to keep up with future system growth. How light boosts bandwidth Our silicon photonics technology is designed to transmit optical signals via single-mode optical fibers, which can support links many tens of kilometers long. Moreover, we have built in the capability to use multiple colors of light, all multiplexed to travel within the same optical fiber, to boost the total data capacity carried. The recently demonstrated silicon photonic chip can combine four wavelengths (all within the telecommunications infrared spectrum), allowing us to transmit four times as much data per fiber. The chip demonstrates transmission and reception of high-speed data at 25 Gb/s over each of these four channels, so within a fully multiplexed design, we’re able to provide 100 Gb/s aggregate bandwidth. A cassette carrying several hundred chips intended for 100 Gb/s transceivers, diced from wafers fabricated with IBM CMOS Integrated Nano-Photonics Technology In addition to the expanded range and bandwidth per fiber, our new photonics technology holds several other advantages over what is available today. Perhaps most importantly, the technology’s manufacturing makes use of conventional silicon microelectronics foundry processes, meaning volume production at low cost. In addition, the entire chip design flow, including simulation, layout, and verification, is enabled by a hardware-verified process design kit, using industry-standard tools. As a result, a high-speed interconnect circuit designer does not require an in-depth knowledge of photonics to build advanced chips with this technology. They can simply follow the standard practices already in place in the CMOS industry. This unified design environment is mirrored by our integrated platform, which allows us to fabricate both the electronic and photonic circuit components on a single silicon chip. Rather than breaking up the electrical and optical functions, we integrated the optical components side by side with sub-100nm CMOS electrical devices. This results in a smaller number of components required to build a transceiver module, as well as a simplified testing and assembly protocol, factors which further contribute to substantial cost reductions. Performance of the fully integrated, wavelength-multiplexed silicon photonics technology demonstrator chip. The eye diagrams illustrate four separate transmitter channels (right) exchanging high-speed data with four receiver channels (left), each running at a rate of 25 Gb/s. While the primary applications for silicon photonics lie within the data center market, driven by Big Data and cloud applications, this technology is also poised to have a large impact within mobile computing. There’s a need for low-cost optical transceivers to shuttle large volumes of data between wireless cellular antennae and their base stations, often located many kilometers away. As the data bandwidth available to mobile users increases generation after generation, the number of individual cells required to support the traffic does the same. Our technology can deliver faster data transfer in higher volume and across larger areas, in order to support the inevitable growth while controlling costs. There has been significant discussion around the 50thanniversary of Moore’s Law and about whether it has reached its end. Silicon photonics fits into that “next switch” conversation. On the processor side, there’s still a fairly consistent trajectory in terms of CMOS technology scaling – down to 10nm, 7nm, and even smaller. The role of our CMOS Integrated Nano-Photonics technology will be to reduce communication bottlenecks inside of systems, and to allow expansion of their capacity for processing huge volumes of data in real time. Kilometer-scale data centers are emerging. Big Data and the Internet of Things are connecting people and information in ways that were unimaginable only a few years ago. IBM’s silicon photonics technology will augment that growth on the ground, into the Cloud, and beyond. Previous Post Next Post More stories ### A new supercomputing-powered weather model may ready us for Exascale In the U.S. alone, extreme weather caused some 297 deaths and $53.5 billion in economic damage in 2016. Globally, natural disasters caused$175 billion in damage. It’s essential for governments, business and people to receive advance warning of wild weather in order to minimize its impact, yet today the information we get is limited. Current […]
2021-09-24 13:07:22
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.22698500752449036, "perplexity": 2302.073204061962}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780057524.58/warc/CC-MAIN-20210924110455-20210924140455-00494.warc.gz"}
https://www.shaalaa.com/question-bank-solutions/pqrs-isosceles-trapezium-l-pq-7-cm-seg-pm-seg-sr-l-sm-3-cm-distance-between-two-parallel-sides-4-cm-find-area-pqrs-area-of-trapezium_81608
Advertisement Remove all ads # ☐ Pqrs is an Isosceles Trapezium L(Pq) = 7 Cm. Seg Pm ⊥ Seg Sr, L(Sm) = 3 Cm, Distance Between Two Parallel Sides is 4 Cm, Find the Area of ☐ Pqrs - Mathematics Sum ☐ PQRS is an isosceles trapezium l(PQ) = 7 cm. seg PM ⊥ seg SR, l(SM) = 3 cm, Distance between two parallel sides is 4 cm, find the area of ☐ PQRS Advertisement Remove all ads #### Solution Draw a perpendicular from Q to line MR. Where it meets the line MR name it point N. MN = PQ = 7 cm In Δ PMS, PM² + SM² = PS² ⇒ 4² + 3² = PS² ⇒ PS² = 16 + 9 = 25 ⇒ PS = 5cm PQRS is an isosceles trapezium so, PS = QR = 5 cm PM = QN = 4 cm So, NR = SM = 3 cm SR = SM + MN + NR = 3 + 7 + 3 = 13 cm Area of trapezium PQRS = 1/2 xx (sum of parallel sides) x height =1/2 xx (7 + 13) x 4 = 40 cm² Is there an error in this question or solution? Advertisement Remove all ads #### APPEARS IN Balbharati Mathematics 8th Standard Maharashtra State Board Chapter 15 Area Practice Set 15.3 | Q 3 | Page 99 Advertisement Remove all ads Advertisement Remove all ads Share Notifications View all notifications Forgot password?
2021-04-14 05:22:32
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5428102612495422, "perplexity": 5629.237822418967}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-17/segments/1618038076819.36/warc/CC-MAIN-20210414034544-20210414064544-00441.warc.gz"}
http://codereview.stackexchange.com/questions/7661/c-unsafe-code-translation
# C# Unsafe code translation? I have some code that I'm using in a standard C# application. I'm sharing the library in a silverlight project that doesn't allow unsafe code. I don't know much at all about unsafe/pointer logic/arethmetic and was wondering if someone could translate the following code snippet so that it will run without /unsafe. I don't care about the performance drop since the code won't be called often on the client. Thanks in advance for any assistance. public static unsafe int GetStableHash(string name) { fixed (char* str = name) { char* chPtr = str; int num = 352654597; int num2 = num; int* numPtr = (int*)chPtr; for (int i = name.Length; i > 0; i -= 4) { num = (((num << 5) + num) + (num >> 27)) ^ numPtr[0]; if (i <= 2) { break; } num2 = (((num2 << 5) + num2) + (num2 >> 27)) ^ numPtr[1]; numPtr += 2; } return (num + (num2 * 1566083941)); } } @JeffMercado Thanks again for your time here guys. I appreciate it. I've tried your updated version and now I get the exception on line num = (((num << 5) + num) + (num >> 27)) ^ BitConverter.ToInt32(bytes, j + 0); The test code I'm using is string[] tmp = new string[] { "Hello", "Isn't that Æegis over there?", "Isn't that something", "11111111111", "111", "112"}; foreach (string s in tmp) { int one = GetStableHash(s); int two = GetStableHashSafe(s); System.Diagnostics.Debug.WriteLine(one + " : " + two + " == " + (one == two).ToString()); } - Questions such as this belongs on Stack Overflow, not here. Fortunately the translation should be simple. –  Jeff Mercado Jan 11 '12 at 6:18 When I posted this question on Stack Overflow, I was told to post it here. Thanks though. –  user1142433 Jan 11 '12 at 6:24 Really? Well I may be wrong but this site is for reviews and optimizations, not to get translations AFAIK. –  Jeff Mercado Jan 11 '12 at 6:29 Understood. It's just hard to know ahead of time sometimes. Thanks again. –  user1142433 Jan 11 '12 at 6:43 This should be pretty much equivalent if I'm not mistaken. public static int GetStableHashSafe(string name) { var bytes = Encoding.Unicode.GetBytes(name + "\0"); //HACK: need null terminator for odd counts int num = 0x15051505; int num2 = num; int j = 0; for (int i = name.Length; i > 0; i -= 4) { num = (((num << 5) + num) + (num >> 27)) ^ BitConverter.ToInt32(bytes, j + 0); if (i <= 2) break; num2 = (((num2 << 5) + num2) + (num2 >> 27)) ^ BitConverter.ToInt32(bytes, j + 4); j += 8; } return (num + (num2 * 0x5D588B65)); } - Hopefully I'm not misunderstanding how the conversion works from string to char*. I'm assuming it's just a simple reinterpretation of the bytes untouched as if the characters are 8-bits in size (as it is in C). –  Jeff Mercado Jan 11 '12 at 6:46 I think that j += 2; should be j += 8; –  Mike Nakis Jan 11 '12 at 6:53 Ah yes, you're right. –  Jeff Mercado Jan 11 '12 at 6:57 Also, I think that int i = bytes.Length; should be int i = name.Length; –  Mike Nakis Jan 11 '12 at 7:02 user: ah yes, I am not sure whether the author of the original unsafe function did this on purpose or by accident, but his code relies on the under-documented fact that strings in C# are zero-terminated so as to facilitate passing them directly to unmanaged functions without having to append a zero character. So, if the string has an odd length, the unmanaged function overshoots and includes the terminator in the calculation. It does not throw an out of range exception because it is unsafe. So, of course you need to emulate that behavior in order to produce the exact same result. Bummer. –  Mike Nakis Jan 11 '12 at 14:24
2014-04-23 09:23:57
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3697580099105835, "perplexity": 4228.021130775454}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-15/segments/1398223202457.0/warc/CC-MAIN-20140423032002-00644-ip-10-147-4-33.ec2.internal.warc.gz"}
https://mathematica.stackexchange.com/questions/161284/solving-an-inequality-with-many-parameters
# Solving an inequality with many parameters [closed] I need to solve the below inequality for x. I am very new to Mathematica. Below syntax does not give me any output. Please help. Reduce[E^[(2 b - T) (x μ - λ)] <= β - (2 a β λ)/(x ν) + (2 a β μ)/ν, x] ## closed as unclear what you're asking by Sektor, Coolwater, gwr, LCarvalho, Henrik SchumacherDec 6 '17 at 8:02 Please clarify your specific problem or add additional details to highlight exactly what you need. As it's currently written, it’s hard to tell exactly what you're asking. See the How to Ask page for help clarifying this question. If this question can be reworded to fit the rules in the help center, please edit the question. • First, syntax is wrong. You can't write E^[]. May be it was cut/paste error. Second, do you know any assumptions on the parameters? Assumptions can help find solutions. – Nasser Dec 4 '17 at 5:54 Reduce[Exp[(2 b - T) (x μ - λ)] <= β - (2 a β λ)/(x ν) + (2 a β μ)/ν, x, Reals] or Reduce[E^((2 b - T) (x μ - λ)) <= β - (2 a β λ)/(x ν) + (2 a β μ)/ν, x, Reals] But Mathematica can't solve this inequality without more information concerning the parameters. When you evaluate either you will get Reduce::nsmet: This system cannot be solved with the methods available to Reduce. • Thank you all for the answers. I have below assumptions on parameters: a<0, 2b-T<0, x μ - λ>0, x>0, actually I need to derive x algebraically, as my theoretical result. Then I can apply appropriate numerical values to all others to find x in my simulations. So isn't there anyway that I can solve this inequality using above information and derive x algebraically? – Babara Peters Dec 4 '17 at 21:33 There is no algebraic solution for x with that inequation. But may be the folowing procedure helps you. According to the given conditions, Exp[...] ranges between 0 and 1 and therefore the right side of the equation is greater zero. So get condtions for the parameters to satisfy at least. red = Reduce[ 0 < β - (2 a β λ)/(x ν) + (2 a β \ μ)/ν && x μ - λ > 0 && x > 0 && a < 0, x, Reals] // LogicalExpand You get 16 different sets of parameter conditions, for example red[[4]]= (-(ν/(2 μ)) == a && x > λ/μ && β > 0 && λ > 0 && μ > 0 && ν > 0) Now insert definite values for the parameters to get exact condition for x Manipulate[ Reduce[Rationalize[ Exp[tbT (x μ - λ)] <= β - (2 a β \ λ)/(x ν) + (2 a β μ)/ν && (x > λ/\ μ && β > 0 && λ > 0 && μ > 0 && ν > 0) /. a -> -(ν/(2 μ)), 0], x] // N, {ν, 1, 5}, {β, 1, 6}, {λ, 1, 4}, {μ, 1, 6}, {tbT, -1, -5}, ContinuousAction -> False]
2019-08-25 01:54:45
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6779069900512695, "perplexity": 2757.7046759664777}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027322160.92/warc/CC-MAIN-20190825000550-20190825022550-00352.warc.gz"}
https://codereview.stackexchange.com/questions/12238/clojure-tictactoe-logic-no-gui
# Clojure TicTacToe (logic, no GUI) I whipped this up last night and was wondering if anyone had any thoughts/comments on it; for example, on: • Code organisation • Coding style • Other improvements • Semantic errors (ns tictactoe.core) (comment " Author: Liam Goodacre Date: 03/06/12 This module allows the creation and progression of tictactoe games. To create a game, use the make-game function: (make-game) A game is a map with the following data: :play-state :playing | :player1 | :player2 | :draw :level-state [:player1|:player2|nil] :scores {:player1 Int, :player2 Int, :draw Int} :turn :player1 | :player2 | :none If the game isn't over, play-state is :playing. If play-state is :player1 or :player2, this indicates that that player won. If play-state is :draw, then the game completed with no winner. level-state is a vector or nine elements. An empty cell has the value of nil. If a player uses a cell, then the cell has that player's id. E.g., if a player1 uses an empty cell, that cell now has the value :player1 scores is a map from ids to integers. It includes score values for draws and for the two players. When the game is over, turn is :none. Otherwise, it describes which player's turn it is. When a new game is created, turn is randomised between the two players. To play a turn, use the play-turn command. For example, assuming 'g' is a predefined game: (play-turn g 2) Will produce a new game state with play progressed by one move, in which the current player used cell 2. Calling new-game clears the level data of a game and randomises the initial player turn. Scores are the only preserved datum. The following are game analysis functions: get-play-state - what state of play are we in? get-turn - whos turn is it? get-scores - get the scores get-level-state - get the state of the level playing? - are we still playing? game-over? - is the game over? valid-move? - is this a valid move? ") (declare ;; Public Functions ; Game Transforming make-game new-game play-turn ; Game Analysing get-play-state get-turn get-scores get-level-state playing? game-over? valid-move? ; Extra get-game-grid) (declare ;; Private Functions ^:private ; Transforming apply-move next-turn recalculate-state ; Analysis calculate-winner try-combination) ;; Utility functions (defmacro ^:private def- " A private version of def. " [& info] (def ^:private ~@info)) (defn- within-range " Determines if a value lies within an inclusive range. " [n x] (fn [v] (and (>= v n) (<= v x)))) ;; Initial data for making new games (def- players [:player1, :player2]) (def- initial-play-state :playing) (def- initial-level-state (vec (repeat 9 nil))) (def- initial-scores {:player1 0, :player2 0, :draw 0}) (defn- random-turn [] (players (int (rand 2)))) (def- winning-combinations [[1 2 3][4 5 6][7 8 9][1 5 9][3 5 7][1 4 7][2 5 8][3 6 9]]) ;; Public Transforming Functions (defn make-game [] " Creates a new game object. " { :play-state initial-play-state :level-state initial-level-state :scores initial-scores :turn (random-turn) }) (defn new-game [g] " Sets up a game object for the next play. " (assoc (make-game) :scores (get-scores g))) (defn play-turn [game move] " Progresses game-play by one move if possible. " (if (and (playing? game) (valid-move? game move)) (-> game (apply-move move) (recalculate-state) (next-turn)) game)) ;; Private Transforming Functions (defn- apply-move [game move] " Progresses game-play by one move. " (assoc game :level-state (assoc (get-level-state game) move (get-turn game)))) (defn- next-turn [game] " Updates which player should play next. If no player is currently active, then do not switch. " (assoc game :turn (case (get-turn game) :none :none :player1 :player2 :player2 :player1))) (defn- recalculate-state [game] " Calculate if there is a winner. If so update the play-state, turn, and scores; to reflect the result. " (let [winner (calculate-winner game) scores (get-scores game)] (if (= winner :none) game (-> game (assoc :play-state winner, :turn :none) (update-in [:scores winner] inc))))) ;; Private Analysis Functions (defn- calculate-winner [game] " Calculates if there is a winner. " (let [state (get-level-state game) matching (partial try-combination state)] (if (some matching winning-combinations) (get-turn game) (if (some nil? state) :none :draw)))) (defn- try-combination [state [one two three]] " Calculates if a winning combination has occurred. " (let [lookup (comp (partial nth state) dec) player (lookup one)] (and (not= player nil) (= player (lookup two) (lookup three))))) ;; Public Analysis Functions (def get-play-state :play-state) (def get-turn :turn) (def get-scores :scores) (def get-level-state :level-state) (defn playing? [game] (= :playing (get-play-state game))) (def game-over? (comp not playing?)) (defn valid-move? [game move] (and ((within-range 0 8) move) (nil? (nth (get-level-state game) move)))) ;; Public Extra Functions (defn get-game-grid [g] " Builds a simple string representation of the current level state. " (clojure.string/join "\n" (map vec (partition 3 (get-level-state g))))) I would put the long (comment ...) as the namespace's docstring. In fact, I would split that in docstrings because you're essentially describing the system by parts in a centralized place, while you could describe it in each part. I would keep the overall "how this game/namespace works/is organized" in namespace's docstring. Declare is meant for forward declarations, not to expose functions (that's what you're doing there, right?) I didn't see it used like that anywhere, but I like the idea! (not= player nil) is just player! In next-turn I would just use a map instead of case (you're essentially mapping, right?) (defn- next-turn [game] (assoc game :turn ((get-turn game) {:none :none ;; I haven't seen spacing to match vertically used too much, but I don't really dislike it :player1 :player2 :player2 :player1})) ;; Or the other way around (the-map (get-turn game)) IMHO, this isn't very idiomatic: (def get-play-state :play-state) ... Aren't you better off using :play-state directly? It offers no encapsulation or maintainability whatsoever (am I right guessing this came from an OOP-like mindset?) And you lose information: the fact that :play-state is a keyword. I guess you did it to expose the analysis functions, right? You're better off mentioning the acceptable keywords in the documentation for your public make-game. I would use either (:play-state game) or (game :play-state) (the second will throw a NullPointerException if game is nil, which isn't bad per se.) This is obviously style but: same goes for Initial data for making new games. I would shove that data directly into the make-game. It'd be already obvious for me that those are the initial states (since you're just bulding a map!) The code is pretty good. I didn't even read the long documentation text and got how it works in a quick glance. I had some troyble understanding how try-combination works (consider making it more readable) but unfortunately I'm a bit in a hurry. I'll come back later and review that too (consider "not easily readable" my review at the moment.) That's my humble opinion. The rest is pretty good. # EDIT @LiamGoodacre that's what I thougth. Clojure favors universal (i.e. reusable) behavior for types instead of "interface tied to type", so it's just the other way around compared to OOP where data and actions are heavily tied. Don't disregard the docstring tips either. Try (doc function-or-namespace) in a REPL to see how important is to have documentation tied to functions/namespaces instead of spread throughout the code or in (comment ...) forms. Documentation generation tools use this metadata too, so it's generally a good idea to do this. I had a hard time grasping try-combination. Consider reworking it a bit, or maybe just add some more info to the docstring like what's [one two three]. The current dosctring is not every useful, I know what the functions does just by reading the name... it tries combinations, and "Calculates if a winning combination has occurred." is just a fancy way of saying it tries combinations. This is just a matter of style, but I dislike heavy use of partial, comp, etc. and would prefer #(nth state (dec %)). I think it's easier to read. I also prefer [a b c] or [x y z] to [one two three]. I had a hard time following you were just destructuring a simple array instead of getting some special values that deserved the long one, two, three names :) Why don't you return the winner in try-combination? Do it like this: (defn- try-combination [state [x y z]] (let [lookup #(nth state (dec %)) player (lookup x)] (and (= player (lookup y) (lookup z)) player))) That way, calculate-winner could return the actual winner and not just the current player, being more correct and reusable IMHO, also allowing you to decouple it from the game and just depending on the current state (e.g. what if you need it later to show the winner in a GUI for recorded games where you only have the final :level-state?) (defn- calculate-winner [state] (let [matching (partial try-combination state) winner (find matching winning-combinations)] (or winner ; Short-circuit logic is cool for these tricks and VERY readable (and (some nil? state) :none) :draw)))) The more decoupled your functions are, the more reusable. Try working with the bare minimum data structures in a single function, instead of using complex maps. I would also use a defrecord to define your game data structure. (defrecord Game [play-state level-state scores turn]) Instance Games like any other Java class, e.g. for make-game : (Game. initial-play-state initial-level-state initial-scores (random-turn)) The good thing: the map interface is still there! IMHO, it offers some benefits: actually encapsulating the tightly coupled structures into an actual object (is there any benefit in the game being a simple map?), the map keys are "documented" as the record args and you can add interfaces to this object. E.g.: you could get rid of get-game-grid and implement it as java.lang.Object's toString, so you could do (str game) and get game's grid (a nice feature IMHO.) You can also add customs protocols to these records to encapsulate the tightly coupled interfaces to your Game object (your Public Analysis Functions, although I would discard most of that.) Please note this is very OOP-y and should be kept to a minimum. Finally: why bother defining functions as private? Screw encapsulation! You'll eventually need these functions outside your namespace for random uninentended tasks (specially if outside people interface with your game, it'd be a cool tic-tac-toe library), so keep these to a minimum (usually helper functions or functions that are tightly coupled with your namespace inner workings.) • Cheers for all the pointers. My main goal with aliasing the keywords (for example :play-state as get-play-state) was to keep a consistant interface with the game: e.g. game-over? is a function, whereas :play-state is a keyword. Having said that, I now see some benefits to reverting these to keywords: play-state is game data, whereas game-over? is a calculated field. – LiamGoodacre Jul 24 '12 at 22:34 • @LiamGoodacre you're welcome :) I did a heavy edit commenting on the issue and adding some more tips after grasping try-combination`. – Álvaro Cuesta Aug 5 '12 at 16:05
2019-06-15 23:56:26
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.34596362709999084, "perplexity": 10656.744792491501}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560627997501.61/warc/CC-MAIN-20190615222657-20190616004657-00423.warc.gz"}
https://gmatclub.com/forum/the-computer-company-registered-a-16-million-net-loss-for-44176.html
Author Message TAGS: ### Hide Tags Senior Manager Joined: 02 Sep 2006 Posts: 254 Followers: 1 Kudos [?]: 62 [0], given: 0 The computer company registered a $16 million net loss for [#permalink] ### Show Tags 06 Apr 2007, 01:21 3 This post was BOOKMARKED 00:00 Difficulty: 15% (low) Question Stats: 71% (01:39) correct 29% (00:45) wrong based on 126 sessions ### HideShow timer Statistics The computer company registered a$16 million net loss for the year, largely because it was profitable only overseas, where much of their profits were used for paying higher taxes, while continuing to lose money money in North America. a) where much of their profits were used for paying higher taxes, while continuing b) where much of their profit went to pay higher taxes, as they continued c) where much of its profit went to pay higher taxes, while it continued d) with much of their profits going to pay higher taxes, while continuing e) with much of the profit used for paying higher taxes, as it continued [Reveal] Spoiler: OA Last edited by WaterFlowsUp on 08 Jan 2015, 00:52, edited 1 time in total. If you have any questions New! VP Joined: 15 Jul 2004 Posts: 1473 Schools: Wharton (R2 - submitted); HBS (R2 - submitted); IIMA (admitted for 1 year PGPX) Followers: 22 Kudos [?]: 178 [0], given: 13 Re: The computer company registered a $16 million net loss for [#permalink] ### Show Tags 06 Apr 2007, 04:56 faifai0714 wrote: The computer company registered a$16 million net loss for the year, largely because it was profitable only overseas, where much of their profits were used for paying higher taxes, while continuing to lose money money in North America. a) where much of their profits were used for paying higher taxes, while continuing --> Pronoun problem. b) where much of their profit went to pay higher taxes, as they continued again pronoun problem c) where much of its profit went to pay higher taxes, while it continued profit went to pay is awkward d) with much of their profits going to pay higher taxes, while continuing Pronoun problem e) with much of the profit used for paying higher taxes, as it continued Right. In C - profit went to pay higher taxes is awkward. I think E is the right answer. VP Joined: 03 Apr 2007 Posts: 1367 Followers: 4 Kudos [?]: 625 [0], given: 10 ### Show Tags 06 Apr 2007, 10:05 C why C ?? - "Its" correctly refers to company i.e singular - "profit" correctly is singular - original has "profits" Manager Joined: 07 Feb 2007 Posts: 212 Followers: 2 Kudos [?]: 13 [1] , given: 0 Re: The computer company registered a $16 million net loss for [#permalink] ### Show Tags 06 Apr 2007, 18:32 1 This post received KUDOS The computer company registered a$16 million net loss for the year, largely because it was profitable only overseas, where much of their profits were used for paying higher taxes, while continuing to lose money money in North America. a) where much of their profits were used for paying higher taxes, while continuing b) where much of their profit went to pay higher taxes, as they continued c) where much of its profit went to pay higher taxes, while it continued--Correct d) with much of their profits going to pay higher taxes, while continuing--Modifier e) with much of the profit used for paying higher taxes, as it continued Intern Joined: 05 Apr 2007 Posts: 9 Followers: 0 Kudos [?]: 0 [0], given: 0 Re: The computer company registered a $16 million net loss for [#permalink] ### Show Tags 06 Apr 2007, 23:39 The computer company registered a$16 million net loss for the year, largely because it was profitable only overseas, where much of their profits were used for paying higher taxes, while continuing to lose money money in North America. ==Parallel construction === a) where much of their profits were used for paying higher taxes, while continuing == company/their don't agree == b) where much of their profit went to pay higher taxes, as they continued == company/their don't agree == c) where much of its profit went to pay higher taxes, while it continued == Correct == d) with much of their profits going to pay higher taxes, while continuing == company/their don't agree == e) with much of the profit used for paying higher taxes, as it continued == awkward construction == Manager Joined: 28 Feb 2007 Posts: 197 Location: California Followers: 3 Kudos [?]: 8 [0], given: 0 Re: The computer company registered a $16 million net loss for [#permalink] ### Show Tags 07 Apr 2007, 01:58 faifai0714 wrote: The computer company registered a$16 million net loss for the year, largely because it was profitable only overseas, where much of their profits were used for paying higher taxes, while continuing to lose money money in North America. a) where much of their profits were used for paying higher taxes, while continuing b) where much of their profit went to pay higher taxes, as they continued c) where much of its profit went to pay higher taxes, while it continued d) with much of their profits going to pay higher taxes, while continuing e) with much of the profit used for paying higher taxes, as it continued A B and D are out for use of 'their' C seems wrong as 'went to pay' is unidiomatic E seems best Senior Manager Joined: 02 Sep 2006 Posts: 254 Followers: 1 Kudos [?]: 62 [0], given: 0 ### Show Tags 10 Apr 2007, 01:37 A,B,D -> 'their' points to overseas. Incorrect since the overseas sites did not pay higher taxes Between C and E, we need 'while' to emphasize that the loss of money was constant. Senior Manager Joined: 03 Dec 2012 Posts: 367 Followers: 0 Kudos [?]: 124 [0], given: 291 ### Show Tags 07 Jun 2015, 04:26 Hello from the GMAT Club VerbalBot! Thanks to another GMAT Club member, I have just discovered this valuable topic, yet it had no discussion for over a year. I am now bumping it up - doing my job. I think you may find it valuable (esp those replies with Kudos). Want to see all other topics I dig out? Follow me (click follow button on profile). You will receive a summary of all topics I bump in your profile area as well as via email. Current Student Joined: 05 Apr 2015 Posts: 450 Followers: 6 Kudos [?]: 180 [0], given: 39 ### Show Tags 19 Sep 2016, 15:01 Hello from the GMAT Club VerbalBot! Thanks to another GMAT Club member, I have just discovered this valuable topic, yet it had no discussion for over a year. I am now bumping it up - doing my job. I think you may find it valuable (esp those replies with Kudos). Want to see all other topics I dig out? Follow me (click follow button on profile). You will receive a summary of all topics I bump in your profile area as well as via email. Re: The computer company registered a $16 million net loss for [#permalink] 19 Sep 2016, 15:01 Similar topics Replies Last post Similar Topics: 4 Biotechnology companies have invested millions of dollars... 7 13 Oct 2013, 10:11 A IT Company recently raised$500 million from several 2 09 Aug 2011, 21:31
2017-02-21 18:10:29
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3003808259963989, "perplexity": 12238.332609986011}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-09/segments/1487501170794.46/warc/CC-MAIN-20170219104610-00266-ip-10-171-10-108.ec2.internal.warc.gz"}
https://raisingthebar.nl/2022/10/24/marginal-probabilities-and-densities/
24 Oct 22 ## Marginal probabilities and densities This is to help everybody, from those who study Basic Statistics up to Advanced Statistics ST2133. ### Discrete case Suppose in a box we have coins and banknotes of only two denominations: $1 and$5 (see Figure 1). Figure 1. Illustration of two variables We pull one out randomly. The division of cash by type (coin or banknote) divides the sample space (shown as a square, lower left picture) with probabilities $p_{c}$ and $p_{b}$ (they sum to one). The division by denomination ($1 or$5) divides the same sample space differently, see the lower right picture, with the probabilities to pull out $1 and$5 equal to $p_{1}$ and $p_{5}$, resp. (they also sum to one). This is summarized in the tables Variable 1: Cash type Prob coin $p_{c}$$p_{c}$ banknote $p_{b}$$p_{b}$ Variable 2: Denomination Prob $1 $p_{1}$$p_{1}$$5 $p_{5}$$p_{5}$ Now we can consider joint events and probabilities (see Figure 2, where the two divisions are combined). Figure 2. Joint probabilities For example, if we pull out a random $item$ it can be a $coin$ and \$1 and the corresponding probability is $P\left(item=coin,\ item\ value=\1\right) =p_{c1}.$ The two divisions of the sample space generate a new division into four parts. Then geometrically it is obvious that we have four identities: Adding over denominations: $p_{c1}+p_{c5}=p_{c},$ $p_{b1}+p_{b5}=p_{b},$ Adding over cash types: $p_{c1}+p_{b1}=p_{1},$ $p_{c5}+p_{b5}=p_{5}.$ Formally, here we use additivity of probability for disjoint events $P\left( A\cup B\right) =P\left( A\right) +P\left( B\right) .$ In words: we can recover own probabilities of variables 1,2 from joint probabilities. ### Generalization Suppose we have two discrete random variables $X,Y$ taking values $x_{1},...,x_{n}$ and $y_{1},...,y_{m},$ resp., and their own probabilities are $P\left( X=x_{i}\right) =p_{i}^{X},$ $P\left(Y=y_{j}\right) =p_{j}^{Y}.$ Denote the joint probabilities $P\left(X=x_{i},Y=y_{j}\right) =p_{ij}.$ Then we have the identities (1) $\sum_{j=1}^mp_{ij}=p_{i}^{X},$ $\sum_{i=1}^np_{ij}=p_{j}^{Y}$ ($n+m$ equations). In words: to obtain the marginal probability of one variable (say, $Y$) sum over the values of the other variable (in this case, $X$). The name marginal probabilities is used for $p_{i}^{X},p_{j}^{Y}$ because in the two-dimensional table they arise as a result of summing table entries along columns or rows and are displayed in the margins. ### Analogs for continuous variables with densities Suppose we have two continuous random variables $X,Y$ and their own densities are $f_{X}$ and $f_{Y}.$ Denote the joint density $f_{X,Y}$. Then replacing in (1) sums by integrals and probabilities by densities we get (2) $\int_R f_{X,Y}\left( x,y\right) dy=f_{X}\left( x\right) ,\ \int_R f_{X,Y}\left( x,y\right) dx=f_{Y}\left( y\right) .$ In words: to obtain one marginal density (say, $f_{Y}$) integrate out the other variable (in this case, $x$).
2023-02-01 09:26:21
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 70, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9281612634658813, "perplexity": 886.6949832197892}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499919.70/warc/CC-MAIN-20230201081311-20230201111311-00877.warc.gz"}
https://physics.stackexchange.com/questions/307994/aerostatics-how-to-calculate-piezometric-head-or-flow-potential-for-a-column
# Aerostatics, how to calculate piezometric head (or flow potential) for a column of gas at rest? [closed] Q:How does one calculate the piezometric head (or flow potential) for a column of gas at rest? Info: From hydrostatics we know that for a liquid continuum at rest or in uniform flow in a gravity field, pressure variations obey: $$\tag{1} -\frac{\partial p}{\partial z}=\rho g=\gamma; \ \ \frac{\partial p}{\partial x}=\frac{\partial p}{\partial y}=0$$ Hence, for a liquid column of constant density ($\rho$), Eqn 1 can be integrated from two points of elevation in the column to obtain: $$\tag{2} z+p/\gamma=\text{const.} \ \ \text{or:} \ \ p_2=p_1-\rho g(z_2 - z_1)$$ which is the equation for hydrostatic pressure distribution. Using the properties of water, I have calculated the hydrostatic pressure distribution assuming the gauge pressure $p_2 = 0$ Pa at a free surface of elevation $z_2=10$ m to an elevation of $z_1=0$ m. The results of which are shown in the figure below, on the left. The quotient $p/\gamma$ in Eqn 2 is called the pressure head and $z$ represents the elevation head, the sum of which is called the piezometric head $\varphi$ (or Hydraulic head), expressed in units of length. For a fluid column at rest the piezometric head is constant everywhere, as shown plotted red in the figure above on the right, using the same water pressure data (given in the figure below). These figures show the components of the piezometric head $\varphi$ or flow potential $\Phi$ (plot on the left) (Note: there are several ways to define a potential. The flow potential, $\Phi$, used here is in units of pressure rather than length). I then considered a static column of nitrogen gas. I used the barometric formula to calculate pressure and density as a function of elevation to try to reproduce the results I found assuming a liquid of constant density, i.e., to calculate a constant piezometric head $\varphi$ or flow potential $\Phi$ for the static column of nitrogen gas. This time, instead of calculating the static pressure distribution from the top of the fluid column, I started from the base ($z_1=0$ m), assuming an absolute pressure of $p_1=101,325$ Pa, and isothermal temperature of 68F (293.15 K), using the barometric formula of the following form: $$\tag{3} p_2 = p_1 \cdot \exp\left(-\frac{Mg}{RT}(z_2 - z_1)\right)$$ Likewise, for density I used $$\tag{4} \rho_2 = \rho_1 \cdot \exp\left(-\frac{Mg}{RT}(z_2 - z_1)\right)$$ where the base density ($\rho_1$) used was 1.165 kg/m3. The results of these calculations, shown below, did not produce a constant piezometric head $\varphi$ or flow potential $\Phi$. I think my error stems from how I am calculating the elevation potential, but I am unsure. Update The issue I was having with the elevation potential, the $\rho g z$ term, was that I was calculating the density of gas at an elevation and then using that (constant) density value for the entire column height. However as we know for gases, density varies with pressure. Therefore, to evaluate the aerostatic/hydrostatic pressure loss/gain, the column of gas should be subdivided into a sufficient number of segments, such that the density in each segment can be assumed to be constant. A numerical integration scheme can be used, such as trapezoidal integration, to sum the contributions of pressure of each subdivision. Alternatively, we can substitute the barometric formula for density, Eqn 4, into our relationship between pressure, density, and elevation, Eqn 1: $\partial p =- \rho g\ \partial z$ and then integrate this equation with respect to elevation, e.g., $$\tag{5} \int_{p_1}^{p(z)} dp=g\rho_1\int_{z_1}^{z}{e^{-a(z-z_1)}}\ dz$$ with $a=Mg/RT$. This integrates to $$\tag{6} p(z)-p_1=\frac{g\rho_1}{a}(1-e^{-a(z-z_1)})$$ See here for details on this integration. Using Equation 6 to calculate the elevation potential produces the expected results as seen in the figure below. However, I am not quite sure how to calculate the pressure head, $p/\rho g$ or $p/\gamma$....thoughts? Update 2 For compressible fluid under isothermal conditions the pressure head ($\Psi$) is defined by $$\tag{7} \Psi=\int_{p_1}^p \frac{dp}{g \rho(p)}$$ Assuming an ideal gas, the relationship between pressure and density is $$\tag{8} \rho = \frac{pM}{RT}$$ Substituting Eqn 8 into Eqn 7, and then performing the integration we have $$\tag{9} \ln(p(z)/p_1)/a$$ where $a=\frac{Mg}{RT}$ Using Eqn 9 to calculate the pressure head (as a function of pressure, which is a function of elevation) produced the results seen in the figure below. The hydrostatic or piezometric head remains constant, which indicates that the column of gas is static, a result I expect. However, I'm not sure I feel OK with the negative pressure head values. Thoughts? ## closed as unclear what you're asking by Jon Custer, heather, Steeven, Mark Mitchison, RococoFeb 16 '17 at 6:04 Please clarify your specific problem or add additional details to highlight exactly what you need. As it's currently written, it’s hard to tell exactly what you're asking. See the How to Ask page for help clarifying this question. If this question can be reworded to fit the rules in the help center, please edit the question. • I honestly have no idea where you are going with this, and what the point of it all is supposed to be. All I can see in the above is a number of trivial rearrangements of terms, amounting to demonstrating a result of the form $g(z)=C+f(z)-f(z)=C$, with $g$ and $f$ various functions and some constant $C$. This is a thoroughly unsurprising result. And, no, a result of that kind indicates exactly nothing other than that you have done your algebra correctly. What is it that you are after here? – Pirx Feb 4 '17 at 17:07 • P.S.: Oh, and of course you get a negative "pressure head" because the pressure drops as you increase altitude. – Pirx Feb 4 '17 at 17:18 Assuming an ideal gas and constant temperature as you do the result is trivial: You get $$\frac{p}{\rho}=R\,T=\mbox{const.},$$ with $R$ the specific gas constant and $T$ the absolute temperature. This is the same result, by the way, which you would obtain by simply combining your equations (3) and (4) above. I will say that I have never seen anyone using this terminology (piezometric, or hydraulic, head) in the atmospheric sciences. I can't see this concept as being practically useful in this area, but I may be wrong. It is also worth noting that the assumption of constant temperature doesn't usually make much sense for atmospheric pressure calculations. Standard atmospheric temperature and pressure distributions are discussed here. • I think for compressible fluid under isothermal conditions the pressure head is defined by: $$\int_{p_o}^p \frac{dp}{g \rho(p)}$$ At this moment, I'm not quite sure how to set this up for spreadsheet calculation. As you can see in my last figure, I was able to calculate a constant Flow Potential using the concepts of pressure and elevation potential. I should be able to do the same in the context of "heads": piezometric, pressure, and elevation. – Armadillo Feb 2 '17 at 18:28 • The result of the integration is simply your equation (3), solved for $\Delta z$. – Pirx Feb 2 '17 at 18:33 • P.S.: I am not sure what the point of your trivial rearrangements of terms is supposed to be in general, and specifically why you are interested in your "flow potential" $\Phi$. As far as I can tell you have defined this as $\Phi=p_0+\int \rho(h)\,g\,h\,\mbox{d}h-\int \rho(h)\,g\,h\,\mbox{d}h=p_0$. What in all the world is this supposed to be good for? – Pirx Feb 2 '17 at 18:46 • Darcy originally expressed fluid flow in terms of "head" (echo2.epfl.ch/VICAIRE/mod_3/chapt_5/main.htm), measured as elevations in manometers (or piezometers). If I am using pressure gauges rather than manometers I should use the rearrangement that allows me to use pressure and calculated elevation potential to calculate the flow potential. Usually, when flowing gas the "elevation head or potential" is neglected. I am trying to solidify my understanding and quantify how much a difference this value is when testing at say 10,000 psi and elevation change of 10 meters. – Armadillo Feb 2 '17 at 18:58 • 10 meters at a base pressure of $10{,}000\,\mbox{psi}$ (!!!)? The gravity-induced pressure change in this case is of the order of $100\,\mbox{Pa}$, or $0.014\,\mbox{psi}$, six orders of magnitude less than your base pressure. I doubt you'll have pressure gauges that could tell the difference. There's a reason why atmospheric pressure changes are neglected at such minuscule elevation differences. – Pirx Feb 2 '17 at 19:03
2019-06-18 03:10:47
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7304860353469849, "perplexity": 430.32866954127724}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560627998605.33/warc/CC-MAIN-20190618023245-20190618045245-00435.warc.gz"}
https://algorithmist.com/index.php/UVa_10838
# UVa 10838 ## Summary This problem is reasonably complex, but an exhaustive memoized search of the state space will run in time. This problem will exercise your ability to do bitwise arithmetic. (In fact, memoization isn't strictly necessary - a simple brute-force search is enough to get accepted on the UVa OJ. Still, a memoized search is much faster.) ## Explanation We are given a 4x4 chess board, consisting of only pawns, and we want to determine which player will win if the game is played optimally as well as how long it will take. Since there are 16 spots on the chess board, and each spot can only be in one of three states (empty, white, black), it is easy to see that ${\displaystyle 3^{16}}$ is an upper bound to the total number of states. ${\displaystyle 3^{16}}$ is just a little bit too much memory to run on the judges machine. However, there are quite a few restrictions on the state space, there are a total of 8 pieces, there are at most 4 white pieces and 4 black pieces, there can only be one white piece on the top row, and one black piece on the bottom row. While it's hard to count the actual size of the state space, it's easy to see that is quite a bit smaller than ${\displaystyle 3^{16}}$. A better upper bound: If we have B black and W white pawns, there are ${\displaystyle 16 \choose B}$ ways to place the black pawns and then ${\displaystyle 16-B \choose W}$ ways to place the white pawns. Summing over all possible B and W we get that there are 2267253 positions with at most 4 pawns of each color. Since the state space is reasonably small, we can simply try to explore it all all and use memoization so that we avoid recomputation of identical states. A particularly nice encoding of the state space uses a pair of integers, one integer for the location of the black pieces, and another for the white pieces. We map the state of row zero column zero to the zeroth bit in the integer, row zero column one to the first bit, and in general row ${\displaystyle r}$ column ${\displaystyle c}$ to the ${\displaystyle 4r+c}$ bit. At each state, we try every possible move and then recursively compute the optimal positions from each move. To each state ${\displaystyle s}$ we associate the optimal number of moves and winner as follows. If the optimal number of moves is ${\displaystyle x}$ and white wins, we associate ${\displaystyle x+1}$, otherwise black wins so we associate ${\displaystyle -(x+1)}$. Then it's easy to find out the optimal move from state ${\displaystyle s}$ as a function of all its substates. The substates are easy to find by simply checking each position to see if it contains a pawn, and if so, if it can capture in either direction or move forward. Keep in mind that if we are going to lose, we want to play as long as possible, and that if we are going to win, we want to play as short as possible. ## Gotcha's Make sure your encoding won't suffer from overflow, it's safe to use 64 bit integers, but you may have some problems with 32 bit ints. ## Implementation • You can use hash_map to do state lookups efficiently. ```2 .ppp .... .PPP .... ...p ...p pP.P ...P ``` ## Output ```white (7) black (2) ```
2019-05-22 06:33:30
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 13, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6063622832298279, "perplexity": 293.79684136824886}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-22/segments/1558232256764.75/warc/CC-MAIN-20190522063112-20190522085112-00011.warc.gz"}
https://mathspace.co/textbooks/syllabuses/Syllabus-447/topics/Topic-8214/subtopics/Subtopic-107740/?activeTab=interactive
Ontario 10 Academic (MPM2D) Distances on the plane (using Pythagorean Thm) ## Interactive practice questions Use the triangle and Pythagoras' theorem to complete the following: a Complete the equations: $AC^2=9$AC2=9 $+$+ $\editable{}$ $AC^2$AC2 $=$= $\editable{}$ b Hence find the length of $AC$AC in simplest radical form. Easy Approx 2 minutes The points $P$P $\left(-3,-2\right)$(3,2), $Q$Q $\left(-3,-4\right)$(3,4) and $R$R $\left(1,-4\right)$(1,4) are the vertices of a right-angled triangle, as shown on the number plane. Consider the interval $AB$AB on the graph. How far is the given point $P$P$\left(-15,8\right)$(15,8) from the origin? ### Outcomes #### 10D.AG2.02 Develop the formula for the length of a line segment, and use this formula to solve problems
2021-09-24 19:19:07
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7035509943962097, "perplexity": 4525.479350072376}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780057564.48/warc/CC-MAIN-20210924171348-20210924201348-00408.warc.gz"}
https://stacks.math.columbia.edu/tag/09X4
Lemma 21.31.12. With $X \in \mathop{\mathrm{Ob}}\nolimits (\textit{LC}_{qc})$ and $a_ X : \mathop{\mathit{Sh}}\nolimits (\textit{LC}_{qc}/X) \to \mathop{\mathit{Sh}}\nolimits (X)$ as above: 1. for an abelian sheaf $\mathcal{F}$ on $X$ we have $H^ n(X, \mathcal{F}) = H^ n_{qc}(X, a_ X^{-1}\mathcal{F})$, 2. for $K \in D^+(X)$ we have $H^ n(X, K) = H^ n_{qc}(X, a_ X^{-1}K)$. For example, if $A$ is an abelian group, then we have $H^ n(X, \underline{A}) = H^ n_{qc}(X, \underline{A})$. In your comment you can use Markdown and LaTeX style mathematics (enclose it like $\pi$). A preview option is available if you wish to see how it works out (just click on the eye in the toolbar).
2019-04-26 04:25:50
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 2, "x-ck12": 0, "texerror": 0, "math_score": 0.9916402697563171, "perplexity": 312.28559977987146}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-18/segments/1555578759182.92/warc/CC-MAIN-20190426033614-20190426055614-00300.warc.gz"}
http://mathhelpforum.com/calculus/114761-volume-under-paraboloid-over-disk-print.html
# Volume under paraboloid and over a disk • Nov 15th 2009, 02:50 PM Volume under paraboloid and over a disk Ok its a simple question really... say that I have to find the volume (using polar coordinates) of the solid under the paraboloid z=x^2+y^2 and above the disk x^2+y^2≤9. My approach would be to find the z value of where the cylinder and paraboloid intersect. Then find the volume of the paraboloid using the value of the plane where it intersects, then subtract it from the volume of the cylinder, this seems right to me. I'm getting z=18 for the point where the cylinder and paraboloid intercept, is this right? For some reason my gut is telling me that its 9 not 18. I simply plugged in the radius of the circle into the paraboloid formula. is there another way that I am missing? • Nov 15th 2009, 03:13 PM Ok after calculating this I got 40.5*pi, is it correct? Is there a way that I can check this? • Nov 15th 2009, 03:18 PM Veraltus Do you mean x^2+y^2=9 or do you mean less than or equal to? • Nov 15th 2009, 03:21 PM Its correct: under the paraboloid z=x^2+y^2 above the disk x^2+y^2≤9 • Nov 15th 2009, 03:36 PM Veraltus I got the same answer as you! Using polar transformations: $x=rcos(\theta)$ and $y=rsin(\theta)$ so we have $z= (rcos(\theta))^2+(rsin(\theta))^2$ $z=r^2$ This turns the integral $\int x^2 + y^2 dxdy$ to $\int r^2 * rdrd\theta$ and using that we have $0 from the disc constraint, and we have $0<\theta<2\pi$ because we want the full circle • Nov 15th 2009, 03:38 PM
2017-01-18 19:30:30
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 8, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9162467122077942, "perplexity": 671.618678793918}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-04/segments/1484560280310.48/warc/CC-MAIN-20170116095120-00356-ip-10-171-10-70.ec2.internal.warc.gz"}
https://zbmath.org/?q=an:1256.32015
# zbMATH — the first resource for mathematics On the ideal triangulation graph of a punctured surface. (English. French summary) Zbl 1256.32015 The authors define the ideal triangulation graph $$T(S)$$ of a hyperbolic surface $$S$$ by letting vertices be (isotopy classes of) ideal triangulations and placing edges between triangulations related by an elementary move (defined to be the operation of replacing an edge of the triangulation by another edge). The mapping class group $$\text{Mod}(S)$$ acts on $$T(S)$$, and thus induces an homomorphism $$\text{Mod}(S) \rightarrow \text{Aut}(T(S))$$. They show that except for two low-complexity examples (punctured torus, thrice-punctured sphere, eliminating these examples is equivalent to the condition $$\chi(S) \leq -2$$), this map is an isomorphism. This is analogous to the situation for the curve complex or the arc complex. However, this graph is not Gromov hyperbolic, in contrast to the curve complex. ##### MSC: 32G15 Moduli of Riemann surfaces, Teichmüller theory (complex-analytic aspects in several variables) 30F10 Compact Riemann surfaces and uniformization Full Text: ##### References: [1] Gromov, M., Essays in Group Theory, edited by S.M. Gersten, Hyperbolic groups, 75-263, (1987), Springer-Verlag · Zbl 0634.20015 [2] Harer, J. L., Stability of the homology of the mapping class groups of orientable surfaces, Annals of Math., 121, 215-249, (1985) · Zbl 0579.57005 [3] Hatcher, A., On triangulations of surfaces, Top. and its Appl., 41, 189-194, (1991) · Zbl 0727.57012 [4] Irmak, E.; Korkmaz, M., Automorphisms of the hatcher-Thurston complex, Isr. J. Math., 162, 183-196, (2007) · Zbl 1149.57032 [5] Irmak, E.; McCarthy, J. D., Injective simplicial maps of the arc complex, Turkish Journal of Mathematics, 33, 1-16, (2009) · Zbl 1206.57018 [6] Ivanov, N. V., 1346, Automorphisms of Teichmüller modular groups, 199-270, (1988), Springer-Verlag, Berlin and New York · Zbl 0657.57004 [7] Ivanov, N. V.; McCarthy, J. D., On injective homomorphisms between Teichmüller modular groups, I. Invent. Math., 135, 2, 425-486, (1999) · Zbl 0978.57014 [8] Korkmaz, M., Automorphisms of complexes of curves on punctured spheres and on punctured tori, Topology and its Applications, 95, 2, 85-111, (1999) · Zbl 0926.57012 [9] Korkmaz, M.; Papadopoulos, A., On the arc and curve complex of a surface · Zbl 1194.57026 [10] Luo, F., Automorphisms of the complex of curves, Topology, 39, 2, 283-298, (2000) · Zbl 0951.32012 [11] Margalit, D., Automorphisms of the pants complex, Duke Math. J., 121, 3, 457-479, (2004) · Zbl 1055.57024 [12] Penner, R. C., The decorated Teichmüller space of punctured surfaces, Communications in Mathematical Physics, 113, 299-339, (1987) · Zbl 0642.32012 This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. It attempts to reflect the references listed in the original paper as accurately as possible without claiming the completeness or perfect precision of the matching.
2021-04-19 00:27:25
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8355928063392639, "perplexity": 1969.20227488513}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-17/segments/1618038862159.64/warc/CC-MAIN-20210418224306-20210419014306-00132.warc.gz"}
http://www.chegg.com/homework-help/questions-and-answers/wageweb-service-hrpdi-provides-compensation-information-170-benchmark-positions-human-reso-q3770191
## Statistics WageWeb is a service of HRPDI and provides compensation information on more than 170 benchmark positions in human resources. The October 2003 posting indicated that labor relation managers earn a mean annual salary of $87,300. Assume that annual salaries are normally distributed and have a standard deviation of$9175. (Give your answers correct to four decimal places.) (a) What is the probability that a randomly selected labor relation manager earned more than $101,200 in 2003? (b) A sample of 15 labor relation managers is taken, and annual salaries are reported. What is the probability that the sample mean annual salary falls between$85,800 and \$88,100? • a) mean=87300 SD=9175 Xbar=101200 Z=(101200-87300)/9175=1.514 Corresponding P is:1-0.93450.0655 b) n=15 Z1=(88100-87300)/9175/sqrt(15)=0.33 Z2=(85800-87300)/9175/sqrt(15)=-0.633 Corresponding P is:0.6393-0.3745=0.2648 Get homework help
2013-05-22 03:10:46
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.21169300377368927, "perplexity": 4293.720094896731}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368701233842/warc/CC-MAIN-20130516104713-00070-ip-10-60-113-184.ec2.internal.warc.gz"}
https://www.merlot.org/merlot/viewMaterial.htm?id=975698
# Material Detail ## Projection onto A Nonnegative Max-Heap This video was recorded at Video Journal of Machine Learning Abstracts - Volume 2. We consider the problem of computing the Euclidean projection of a vector of length $p$ onto a non-negative max-heap - an ordered tree where the values of the nodes are all nonnegative and the value of any parent node is no less than the value(s) of its child node(s). This Euclidean projection plays a building block role in the optimization problem with a non-negative max-heap constraint. Such a constraint is desirable when the features follow an ordered tree structure, that is, a given feature is selected for the given regression/classification task only if its parent node is selected. In this paper, we show that such Euclidean projection problem admits an analytical solution and we develop a top-down algorithm where the key operation is to find the so-called \emph{maximal root-tree} of the subtree rooted at each node. A naive approach for finding the maximal root-tree is to enumerate all the possible root-trees, which, however, does not scale well. We reveal several important properties of the maximal root-tree, based on which we design a bottom-up algorithm with merge for efficiently finding the maximal root-tree. The proposed algorithm has a (worst-case) linear time complexity for a sequential list, and $O(p^2)$ for a general tree. We report simulation results showing the effectiveness of the max-heap for regression with an ordered tree structure. Empirical results show that the proposed algorithm has an expected linear time complexity for many special cases including a sequential list, a full binary tree, and a tree with depth 1. #### Quality • User Rating • Learning Exercises • Bookmark Collections • Course ePortfolios • Accessibility Info
2022-05-21 12:05:47
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.599247395992279, "perplexity": 535.9188623953046}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652662539101.40/warc/CC-MAIN-20220521112022-20220521142022-00573.warc.gz"}
https://brilliant.org/problems/how-do-i-even-pronounce-this-word/
How do I even pronounce this word? Like the Fibonacci sequence and Tribonacci sequence, define the Elevbonacci sequence such that its $$n^\text{th}$$ term, $$E_n$$ is the sum of the previous eleven terms with initial terms $$E_0 = E_1 = E_2 = \ldots = E_9 = 0, E_{10} = 1$$. Define $$\displaystyle R = \lim_{n \to \infty} \frac { E_{n+1}}{E_n}$$. Let $$f$$ denote the least degree monic polynomial with integer coefficients such that it has root $$R$$. Evaluate $\sum_{f(r_{i}) = 0} r_{i}^{\deg(f)}$ where $$\deg(f)$$ is the degree of $$f$$. Bonus: can you generalize this? ×
2017-10-22 14:04:49
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9531142711639404, "perplexity": 414.6048370377938}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187825264.94/warc/CC-MAIN-20171022132026-20171022152026-00697.warc.gz"}
https://radar.inria.fr/rapportsactivite/RA2019/hycomes/uid51.html
New Software and Platforms Bilateral Contracts and Grants with Industry Partnerships and Cooperations Bibliography PDF e-Pub ## Section: New Results ### Mathematical Foundations of Physical Systems Modeling Languages Participants : Albert Benveniste, Benoît Caillaud, Mathias Malandain. Modern modeling languages for general physical systems, such as Modelica or Simscape, rely on Differential Algebraic Equations (DAE), i.e., constraints of the form $f\left(\stackrel{˙}{x},x,u\right)=0$. This facilitates modeling from first principles of the physics. This year we completed the development of the mathematical theory needed to sound, on solid mathematical bases, the design of compilers and tools for DAE based physical modeling languages. Unlike Ordinary Differential Equations (ODE, of the form $\stackrel{˙}{x}=g\left(x,u\right)$), DAE exhibit subtle issues because of the notion of differentiation index and related latent equations—ODE are DAE of index zero for which no latent equation needs to be considered. Prior to generating execution code and calling solvers, the compilation of such languages requires a nontrivial structural analysis step that reduces the differentiation index to a level acceptable by DAE solvers. Multimode DAE systems, having multiple modes with mode-dependent dynamics and state-dependent mode switching, are much harder to deal with. The main difficulty is the handling of the events of mode change. Unfortunately, the large literature devoted to the numerical analysis of DAEs does not cover the multimode case, typically saying nothing about mode changes. This lack of foundations causes numerous difficulties to the existing modeling tools. Some models are well handled, others are not, with no clear boundary between the two classes. Basically, no tool exists that performs a correct structural analysis taking multiple modes and mode changes into account. In our work, we developed a comprehensive mathematical approach supporting compilation and code generation for this class of languages. Its core is the structural analysis of multimode DAE systems, taking both multiple modes and mode changes into account. As a byproduct of this structural analysis, we propose well sound criteria for accepting or rejecting models at compile time. For our mathematical development, we rely on nonstandard analysis, which allows us to cast hybrid systems dynamics to discrete time dynamics with infinitesimal step size, thus providing a uniform framework for handling both continuous dynamics and mode change events. A big comprehensive document has been written, which will be finalized and submitted next year.
2023-02-03 03:44:27
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 2, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7270961403846741, "perplexity": 1605.0674826667134}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764500042.8/warc/CC-MAIN-20230203024018-20230203054018-00439.warc.gz"}
http://mathoverflow.net/revisions/25949/list
The word "true" can, however, be defined mathematically. Truth is a property of sentences. If you have defined a formal language $L$, such as the first-order language of arithmetic, then you can define a sentence $S$ in $L$ to be true if and only if $S$ holds of the natural numbers. So for example the sentence $\exists x: x > 0$ is true because there does indeed exist a natural number greater than 0. Here it is important to note that true is not the same as provable. The formal sentence corresponding to the twin prime conjecture (which I won't bother writing out here) is true if and only if there are infinitely many twin primes, and it doesn't matter that we have no idea how to prove or disprove the conjecture. Now, how can we have true but unprovable statements? And if we had one how would we know? Joel David Hamkins explained this well, but in brief, "unprovable" is always with respect to some set of axioms. Therefore it is possible for some statement to be true but unprovable from some particular set of axioms $A$. In order to know that it's true, of course, we still have to prove it, but that will be a proof from some other set of axioms besides $A$.
2013-05-22 14:41:55
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8359447717666626, "perplexity": 80.32485201836239}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368701852492/warc/CC-MAIN-20130516105732-00026-ip-10-60-113-184.ec2.internal.warc.gz"}
https://tex.stackexchange.com/questions/480306/how-to-make-table-with-column-lines-only-in-the-first-row
# How to make table with column lines only in the first row How do I draw my table with the column lines only in the first row? Here's a MWE of my code: \documentclass[12pt,a4paper]{article} \usepackage[utf8]{inputenc} \usepackage{float} \usepackage[table]{xcolor} \begin{document} \begin{table}[H] \centering \caption{Data of subways in Fortal city} \rowcolors{1}{gray!20}{white} \begin{tabular}{c|c|c|c} \hline Line & Extension & Stations & Vehicle type \\ \hline South & 24,1 km & 19 & Electric \\ West & 19,5 km & 10 & VLT a diesel \\ Parangaba-Mucuripe & 13,2 km & 10 & VLT a diesel \\ East & 7,3 km & 11 & VLT a diesel \\ \hline \end{tabular} \end{table} \end{document} What I'd like to achieve: • \multicolumn{1}{c|}{Line} ... – John Kormylo Mar 19 '19 at 15:46 • – user156344 Mar 19 '19 at 15:48 • @JohnKormylo Would you like to add an answer? – CarLaTeX Mar 20 '19 at 6:09 • @CarLaTeX I'll post my answer with the help of them. – Levy Mar 20 '19 at 15:39 It's an easy trick using the multicol package. I removed the columns in my tabular agument but I drew them back, only in the header, using multicolumn{1}{c|}{text} Here's the full working code: \documentclass[12pt,a4paper]{article} \usepackage[utf8]{inputenc} \usepackage{float} \usepackage[table]{xcolor} \usepackage{multicol} \begin{document} \begin{table}[H] \centering \caption{Data of subways in Fortal city} \rowcolors{1}{gray!20}{white} \begin{tabular}{cccc} \hline \multicolumn{1}{c|}{Line} & \multicolumn{1}{c|}{Extension} & \multicolumn{1}{c|}{Stations} & Vehicle type \\ \hline South & 24,1 km & 19 & Electric \\ West & 19,5 km & 10 & VLT a diesel \\ Parangaba-Mucuripe & 13,2 km & 10 & VLT a diesel \\ East & 7,3 km & 11 & VLT a diesel \\ \hline \end{tabular} \end{table} \end{document} Notice that for this particular problem I don't need to draw a column c| in the last column. Otherwise it'd turned into a border, which I don't want to happen. Furthermore, to avoid always writing multicolumn{1}{c|}{text} I defined a new command \topr{} as \newcommand{\topr}[1]{\multicolumn{1}{c|}{#1}} which spare me to write fixed values. So the improved code turned out like this: \documentclass[12pt,a4paper]{article} \usepackage[utf8]{inputenc} \usepackage{float} \usepackage[table]{xcolor} \usepackage{multicol} \newcommand{\topr}[1]{\multicolumn{1}{c|}{#1}} \begin{document} \begin{table}[H] \centering \caption{Data of subways in Fortal city} \rowcolors{1}{gray!20}{white} \begin{tabular}{cccc} \hline \topr{Line} & \topr{Extension} & \topr{Stations} & Vehicle type \\ \hline South & 24,1 km & 19 & Electric \\ West & 19,5 km & 10 & VLT a diesel \\ Parangaba-Mucuripe & 13,2 km & 10 & VLT a diesel \\ East & 7,3 km & 11 & VLT a diesel \\ \hline \end{tabular} \end{table} \end{document} ## calstable Here is your table type setted using the package cals. As you will find, all rules are visible, and there are no white lines between rules and colours. However, I wold recommend typesetting a classic booktable tabular, without vertical rules, and without any gray shading, see below. I have defined some shortcuts to reduce coding. Compared to your original, I have removed side bearing (the left padding in first column and the right padding in the last column, which his customary for academic tables. In addition, I have right aligned the column with figures, but moved the figures to the centre of the column by 1.25em. I also moved the "km" to the column heading. If you prefer one line headings, just increase the column with by some pt. \documentclass[12pt, a4paper]{article} \usepackage{cals, url} \usepackage{xcolor, caption} \newcommand{\rl}{\hspace*{1.25em}} \begin{document} \begin{table*} \caption{Data of subways in Fortal city} %\footnotesize % Not necessary with too tiny font. Eventually, use small if you prefer smaller font in the tables. \small\sffamily \begin{calstable}[c] % Defining columns relative to each other and relative to the margins \colwidths{{\dimexpr(\columnwidth)/3\relax} {\dimexpr(\columnwidth)/6\relax} {\dimexpr(\columnwidth)/6\relax} {\dimexpr(\columnwidth)/4\relax} } % The tabular fills the text area if the divisors for all columns were 4 % Set up the tabular \makeatletter \def\cals@framers@width{0.8pt} % Outside frame rules, reduce if the rule is too heavy \def\cals@framecs@width{0pt} \def\cals@bodyrs@width{0.6pt} \def\cals@cs@width{0.4pt} % Inside rules, reduce if the rule is too heavy \def\cals@rs@width{0.4pt} \def\cals@bgcolor{} \def\lb{\ifx\cals@borderL\relax % Left border switch (off-on) \def\cals@borderL{0pt} \else \let\cals@borderL\relax\fi} \def\rb{\ifx\cals@borderR\relax % Right border switch (off-on) \def\cals@borderR{0pt} \else \let\cals@borderR\relax\fi} \def\bb{\ifx\cals@borderB\relax % Bottom border switch (off-on) \def\cals@borderB{0pt} \else \let\cals@borderB\relax\fi} \def\gray{\ifx\cals@bgcolor\empty % "Switch" to turn on and off colour \def\cals@bgcolor{gray!20} \else \def\cals@bgcolor{} \fi} % R1H1 \brow \gray\lp\alignL\cell{\vfil Line}\lp \alignC\cell{\vfil Extension (km)} \alignC\cell{\vfil Stations} \rp\alignR\cell{\vfil Vehicle type}\rp \gray \erow \mdseries% } \tfoot{\lastrule\strut} %R2H1 \brow \lb\lp\bb\alignL\cell{South}\lp \alignR\cell{24,1\rl} \alignR\cell{19\rl} \rp\alignR\cell{Electric}\rp \erow %R3H2 \brow \gray\lp\alignL\cell{West}\lp \alignR\cell{19,5\rl} \alignR\cell{10\rl} \rp\alignR\cell{VLT a diesel}\rp\gray \erow %R4B3 \brow \lp\alignL\cell{Parangaba-Mucuripe}\lp \alignR\cell{13,2\rl} \alignR\cell{10\rl} \rp\alignR\cell{VLT a diesel}\rp\bb \erow %R7N1 \brow \gray\lp\cell{ East}\lp \alignR\cell{7,3\rl} \alignR\cell{11\rl} \rp\alignR\cell{VLT a diesel}\gray\rp \erow \makeatletter \end{calstable}\par \end{table*} \end{document} ## Classic booktable style tabular Here I have modified @Levy's code, but made several changes: 1. Using tabulary for best fitting column with 2. Used booktabs for correctly spaced rules 3. Removed side bearing (@{}) 5. Columns and headers similar alignment \documentclass[12pt,a4paper]{article} \usepackage{booktabs, caption, tabulary} \begin{document} \begin{table}[!tb] \small\sffamily\centering \caption{Data of subways in Fortal city} \begin{tabulary}{\linewidth}{@{}LRRC@{}} \toprule \textsc{line} & \textsc{extension\linebreak(km)} & \textsc{stations} & \textsc{vehicle type}\\ \midrule South & 24,1 & 19 & Electric \\ West & 19,5 & 10 & VLT a diesel \\ Parangaba-Mucuripe & 13,2 & 10 & VLT a diesel \\ East & 7,3 & 11 & VLT a diesel \\ \bottomrule \end{tabulary} \end{table} \end{document}
2020-01-24 01:24:08
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6263752579689026, "perplexity": 7887.0213934368385}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-05/segments/1579250614880.58/warc/CC-MAIN-20200124011048-20200124040048-00027.warc.gz"}
https://www.semanticscholar.org/paper/Discovery-of-Enhanced-Germanium-Abundances-in-with-Sterling-Dinerstein/e4d00acf94b98e30be75215e70313a554baac487
# Discovery of Enhanced Germanium Abundances in Planetary Nebulae with FUSE @article{Sterling2002DiscoveryOE, title={Discovery of Enhanced Germanium Abundances in Planetary Nebulae with FUSE}, author={N. C. Sterling and Harriet L. Dinerstein and Charles W. Bowers}, journal={arXiv: Astrophysics}, year={2002} } • Published 2002 • Physics • arXiv: Astrophysics We report the discovery of Ge III $\lambda$1088.46 in the planetary nebulae (PNe) SwSt 1, BD+30$^{\rm o}$3639, NGC 3132, and IC 4593, observed with the Far Ultraviolet Spectroscopic Explorer. This is the first astronomical detection of this line and the first measurement of Ge (Z = 32) in PNe. We estimate Ge abundances using S and Fe as reference elements, for a range of assumptions about gas-phase depletions. The results indicate that Ge, which is synthesized in the initial steps of the s… Expand 5 Citations The Production of Germanium in Asymptotic Giant Branch stars • Physics • 2006 Observations of planetary nebulae by Sterling, Dinerstein and Bowers have revealed enhancements in the element Germanium (Ge) by factors of 3 to 10 compared to solar. Germanium can be produced byExpand Physical properties of the fluorine and neutron-capture element-rich PN Jonckheere 900 • Physics • 2019 We performed detailed spectroscopic analyses of a young C-rich planetary nebula (PN) Jonckheere900 (J900) in order to characterise the properties of the central star and nebula. Of the derived 17Expand Germanium, Arsenic, and Selenium Abundances in Metal-Poor Stars The elements germanium (Ge, Z = 32), arsenic (As, Z = 33), and selenium (Se, Z = 34) span the transition from charged-particle or explosive synthesis of the iron-group elements to neutron-captureExpand Accurate determination of energy levels, hyperfine structure constants, lifetimes and dipole polarizabilities of triply ionized tin isotopes • Physics • 2020 We have investigated energies, magnetic dipole hyperfine structure constants ($A_{hyf}$) and electric dipole (E1) matrix elements of a number of low-lying states of the triply ionized tin (Sn$^{3+}$)Expand Physical Conditions and Chemical Abundances in Photoionized Nebulae from Optical Spectra This chapter presents a review on the latest advances in the computation of physical conditions and chemical abundances of elements present in photoionized gas (H ii regions and planetary nebulae).Expand #### References The origin of the elements. The development of the modern understanding of the origin of the chemical elements is reviewed. Nucleosynthesis in stars and in the early Universe is treated in this 1978 Nobel lecture. Reference isExpand
2021-10-21 19:06:31
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6903427839279175, "perplexity": 7824.682332666532}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-43/segments/1634323585439.59/warc/CC-MAIN-20211021164535-20211021194535-00656.warc.gz"}
https://codedump.io/share/GmYHvuY6a1ow/1/mathjax-works-in-desktop-browsers-but-only-when-faking-user-agent-works-on-mobile-browsers
Marcus M&#252;ller - 1 year ago 75 Javascript Question MathJax: works in desktop browsers, but only when faking user-agent; works on mobile browsers http://gnuradio.org/blog/filtering-time-series-data__elemental-building-blocks/ , which should contain a rendering of my LaTeX formulas, exactly like: It does the same (works!) when I configure my Desktop's Firefox to pretend its user-agent was that of a Firefox running on Android Sadly, without forcing my firefox to claim its user-agent string is that of an IPhone 3, the formula just never gets rendered; MathJax shows it's typical "initialization messages", but long before these appear, the "raw" LaTeX formula just disappears and never comes back rendered: Now, the peculiar thing is: It doesn't work with Chrome under Linux, either, and I've got a friend who reports the same problem with Safari/OS X but with Firefox on an Android device, it works like a charm. Same goes for Chrome on Android. So, I tried debugging things, but I haven't gotten beyond the following: • Network requests seem to be identical, and so do the responses, no matter what user-agent I use on my desktop PC, aside from the fact that when things work, there's some traffic with cdn.mathjax.org (which is expected, as it loads webfonts) • Putting a render job into Mathjax Hub 's Queue doesn't help either. Not being a web dev at all, this is pretty much how far I've gotten. So both hints about what to poke at / use to debug this or definite answers solving my problem are equally welcome. The problem is that you page is loading code that modifies script elements in the page (http://gnuradio.org/cdn-cgi/nexp/dok3v=1613a3a185/cloudflare/rocket.js, which is titled cloudflarejs-rocketloader-0.11.5 within the file. It may be that cloudflare is adding this automatically without your knowledge, but the result is that the script elements in your page are no longer the standard script DOM elements implemented by the browser, but are a replacement implemented by this script. Unfortunately, the behavior of the replacement is not identical to the original script elements, and that is causing MathJax to not be able to read the math (which is stores in script elements within the page). You may be able to overcome that by adding <script> MathJax = { AuthorInit: function () { MathJax.HTML.getScript = function (node) {return node.textContent} } }; </script>
2018-01-23 04:35:15
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4294168949127197, "perplexity": 2653.718614097424}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-05/segments/1516084891706.88/warc/CC-MAIN-20180123032443-20180123052443-00240.warc.gz"}
https://studyadda.com/solved-papers/neet-solved-paper-2016-phase-i_q166/435/288751
• # question_answer Water vapour comes out from the plant leaf through the stomatal opening. Through the same stomatalopening carbon dioxide diffuses into the plant during photosynthesis. Reason out the above statements using oneof following options: A)  Both processes cannot happen simultaneouslyB)  Both processes can happen together because the diffusion coefficient of water and $C{{O}_{2}}$ is differentC)  The above processes happen only during night timeD)  One process occurs during day time, and the other at night Water vapour comes out and $C{{O}_{2}}$ diffuses simultaneously through stomatal opening because diffusion coefficient of water and $C{{O}_{2}}$ is different and it moves along its concentration gradient.
2020-02-20 04:41:13
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3036317229270935, "perplexity": 3419.0608052873895}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875144637.88/warc/CC-MAIN-20200220035657-20200220065657-00412.warc.gz"}
http://secolinsky.com/blogs/gre_problems_13_14,2/
The more evidence we have to support our inferences and the fewer interpretaions possible, the more we can trust their accuracy GRE Math Practice Test Problems By Maurice Ticas Studying for the exam I've encountered two problems. The first is stated as follows: If $$f$$ is a continuously differentiable real-valued function defined on the open interval $$(-1,4)$$ such that $$f(3)=5$$ and $$f'(x)\geq-1$$ for all $$x$$, what's the greatest value of $$f(0)$$? The second I thought was difficult too and worth sharing: Suppose $$g$$ is a continuous real-valued function such that $$3x^5+96=\int_c^x g(t) dt$$ for each $$x \in \mathbb{R}$$, where $$c$$ is a constant. What is the value of $$c$$? What approach do you have to answer the two problems? There are 0 comments. No more comments are allowed.
2019-02-16 11:21:21
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7983552813529968, "perplexity": 197.2970094055749}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-09/segments/1550247480272.15/warc/CC-MAIN-20190216105514-20190216131514-00244.warc.gz"}
https://www.rdocumentation.org/packages/spatstat/versions/1.59-0/topics/Gcom
# Gcom 0th Percentile ##### Model Compensator of Nearest Neighbour Function Given a point process model fitted to a point pattern dataset, this function computes the compensator of the nearest neighbour distance distribution function $G$ based on the fitted model (as well as the usual nonparametric estimates of $G$ based on the data alone). Comparison between the nonparametric and model-compensated $G$ functions serves as a diagnostic for the model. Keywords models, spatial ##### Usage Gcom(object, r = NULL, breaks = NULL, ..., correction = c("border", "Hanisch"), conditional = !is.poisson(object), restrict=FALSE, model=NULL, trend = ~1, interaction = Poisson(), rbord = reach(interaction), ppmcorrection="border", truecoef = NULL, hi.res = NULL) ##### Arguments object Object to be analysed. Either a fitted point process model (object of class "ppm") or a point pattern (object of class "ppp") or quadrature scheme (object of class "quad"). r Optional. Vector of values of the argument $r$ at which the function $G(r)$ should be computed. This argument is usually not specified. There is a sensible default. breaks This argument is for internal use only. correction Edge correction(s) to be employed in calculating the compensator. Options are "border", "Hanisch" and "best". Alternatively correction="all" selects all options. conditional Optional. Logical value indicating whether to compute the estimates for the conditional case. See Details. restrict Logical value indicating whether to compute the restriction estimator (restrict=TRUE) or the reweighting estimator (restrict=FALSE, the default). Applies only if conditional=TRUE. See Details. model Optional. A fitted point process model (object of class "ppm") to be re-fitted to the data using update.ppm, if object is a point pattern. Overrides the arguments trend,interaction,rbord,ppmcorrection. trend,interaction,rbord Optional. Arguments passed to ppm to fit a point process model to the data, if object is a point pattern. See ppm for details. Extra arguments passed to ppm. ppmcorrection The correction argument to ppm. truecoef Optional. Numeric vector. If present, this will be treated as if it were the true coefficient vector of the point process model, in calculating the diagnostic. Incompatible with hi.res. hi.res Optional. List of parameters passed to quadscheme. If this argument is present, the model will be re-fitted at high resolution as specified by these parameters. The coefficients of the resulting fitted model will be taken as the true coefficients. Then the diagnostic will be computed for the default quadrature scheme, but using the high resolution coefficients. ##### Details This command provides a diagnostic for the goodness-of-fit of a point process model fitted to a point pattern dataset. It computes different estimates of the nearest neighbour distance distribution function $G$ of the dataset, which should be approximately equal if the model is a good fit to the data. The first argument, object, is usually a fitted point process model (object of class "ppm"), obtained from the model-fitting function ppm. For convenience, object can also be a point pattern (object of class "ppp"). In that case, a point process model will be fitted to it, by calling ppm using the arguments trend (for the first order trend), interaction (for the interpoint interaction) and rbord (for the erosion distance in the border correction for the pseudolikelihood). See ppm for details of these arguments. The algorithm first extracts the original point pattern dataset (to which the model was fitted) and computes the standard nonparametric estimates of the $G$ function. It then also computes the model-compensated $G$ function. The different functions are returned as columns in a data frame (of class "fv"). The interpretation of the columns is as follows (ignoring edge corrections): bord: the nonparametric border-correction estimate of $G(r)$, $$\hat G(r) = \frac{\sum_i I\{ d_i \le r\} I\{ b_i > r \}}{\sum_i I\{ b_i > r\}}$$ where $d_i$ is the distance from the $i$-th data point to its nearest neighbour, and $b_i$ is the distance from the $i$-th data point to the boundary of the window $W$. bcom: the model compensator of the border-correction estimate $${\bf C}\, \hat G(r) = \frac{\int \lambda(u,x) I\{ b(u) > r\} I\{ d(u,x) \le r\}}{ 1 + \sum_i I\{ b_i > r\} }$$ where $\lambda(u,x)$ denotes the conditional intensity of the model at the location $u$, and $d(u,x)$ denotes the distance from $u$ to the nearest point in $x$, while $b(u)$ denotes the distance from $u$ to the boundary of the window$W$. han: the nonparametric Hanisch estimate of $G(r)$ $$\hat G(r) = \frac{D(r)}{D(\infty)}$$ where $$D(r) = \sum_i \frac{ I\{x_i \in W_{\ominus d_i}\} I\{d_i \le r\} }{ \mbox{area}(W_{\ominus d_i}) }$$ in which $W_{\ominus r}$ denotes the erosion of the window $W$ by a distance $r$. hcom: the corresponding model-compensated function $${\bf C} \, G(r) = \int_W \frac{ \lambda(u,x) I(u \in W_{\ominus d(u)}) I(d(u) \le r) }{ \hat D(\infty) \mbox{area}(W_{\ominus d(u)}) + 1 }$$ where $d(u) = d(u, x)$ is the (empty space') distance from location $u$ to the nearest point of $x$. If the fitted model is a Poisson point process, then the formulae above are exactly what is computed. If the fitted model is not Poisson, the formulae above are modified slightly to handle edge effects. The modification is determined by the arguments conditional and restrict. The value of conditional defaults to FALSE for Poisson models and TRUE for non-Poisson models. If conditional=FALSE then the formulae above are not modified. If conditional=TRUE, then the algorithm calculates the restriction estimator if restrict=TRUE, and calculates the reweighting estimator if restrict=FALSE. See Appendix E of Baddeley, Rubak and Moller (2011). See also spatstat.options('eroded.intensity'). Thus, by default, the reweighting estimator is computed for non-Poisson models. The border-corrected and Hanisch-corrected estimates of $G(r)$ are approximately unbiased estimates of the $G$-function, assuming the point process is stationary. The model-compensated functions are unbiased estimates of the mean value of the corresponding nonparametric estimate, assuming the model is true. Thus, if the model is a good fit, the mean value of the difference between the nonparametric and model-compensated estimates is approximately zero. To compute the difference between the nonparametric and model-compensated functions, use Gres. ##### Value A function value table (object of class "fv"), essentially a data frame of function values. There is a plot method for this class. See fv.object. ##### References Baddeley, A., Rubak, E. and Moller, J. (2011) Score, pseudo-score and residual diagnostics for spatial point process models. Statistical Science 26, 613--646. Related functions: Gest, Gres. Alternative functions: Kcom, psstA, psstG, psst. Model fitting: ppm. • Gcom ##### Examples # NOT RUN { data(cells) fit0 <- ppm(cells, ~1) # uniform Poisson G0 <- Gcom(fit0) G0 plot(G0) # uniform Poisson is clearly not correct # Hanisch estimates only plot(Gcom(fit0), cbind(han, hcom) ~ r) fit1 <- ppm(cells, ~1, Strauss(0.08)) plot(Gcom(fit1), cbind(han, hcom) ~ r) `
2019-07-16 10:01:36
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7053768038749695, "perplexity": 1715.816638677708}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-30/segments/1563195524522.18/warc/CC-MAIN-20190716095720-20190716121720-00315.warc.gz"}
https://www.authorea.com/users/8363/articles/8798/_show_article
# The $$\gamma$$-ray Milky Way above 10 GeV: Distinguishing Sources from Diffuse Emission Abstract One of the most prominent features of the gamma-ray sky is th
2017-09-25 06:21:53
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8855258226394653, "perplexity": 5840.353304413851}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-39/segments/1505818690340.48/warc/CC-MAIN-20170925055211-20170925075211-00642.warc.gz"}
https://quantumcomputing.stackexchange.com/questions/6266/quantum-operation-involving-permutation
# Quantum operation involving permutation Suppose I have the state $$\frac{1}{2^l/2}\sum_{i=0}^{2^l-1}|0\rangle^{\otimes q}\otimes |i\rangle^{\otimes}|0\rangle_i^{\otimes l}$$. I perform some unitary transformation between the registers $$|i\rangle^{\otimes}$$ and $$|0\rangle_i^{\otimes l}$$. That unitary operation is just the permutation, i.e. for each $$|i\rangle^{\otimes l}|$$ $$0\rangle_i^{\otimes l}$$ stores the permutation of $$|i\rangle^{\otimes l}$$. Originally each $$|0\rangle^{\otimes q}$$ corresponded to each $$|i\rangle^{\otimes}$$. But now I want that $$|0\rangle^{\otimes q}$$ be associated to the permuted value stored in $$|0\rangle_i^{\otimes l}$$ correspinding to $$|i\rangle^{\otimes l}$$. For example initially I have the pairs $$(5,0), (10,1), (45,2), (59,3)$$ Now I perform the permutation on the second coordinate to obtain $$0\to 1$$, $$1\to 3$$, $$2\to 0$$, $$3 \to 2$$. So now I want pairs to be $$(5,1), (10,3), (45,0), (59,2)$$. How can this be obtained? Can someone help me out? Edit: I give what i think of the solution. Consider an example of $$4$$ states as ignore the normalizing factor. $$|A\rangle= |5\rangle |0\rangle|0\rangle+|10\rangle |1\rangle|0\rangle+|45\rangle |2\rangle|0\rangle+|59\rangle |3\rangle|0\rangle$$ Now i apply the permutation on the second register to store the result in the third register to get $$|B\rangle= |5\rangle |0\rangle|1\rangle+|10\rangle |1\rangle|3\rangle+|45\rangle |2\rangle|0\rangle+|59\rangle |3\rangle|2\rangle$$ Now i swap the second and third qubits to get $$|C\rangle= |5\rangle |1\rangle|0\rangle+|10\rangle |3\rangle|1\rangle+|45\rangle |0\rangle|2\rangle+|59\rangle |2\rangle|3\rangle$$ Now since the third register is just the Hadamard on $$2$$ qubits, so i apply Hadamard on the third register of state $$|C\rangle$$ to get $$|D\rangle= |5\rangle |1\rangle|0\rangle+|10\rangle |3\rangle|0\rangle+|45\rangle |0\rangle|0\rangle+|59\rangle |2\rangle|0\rangle$$ Is this reasoning correct? • Won't applying the permutation on the second register alone suffice? – Mahathi Vempati Jun 3 at 4:14
2019-09-15 17:53:40
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 24, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8649242520332336, "perplexity": 372.4599855355505}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-39/segments/1568514572235.63/warc/CC-MAIN-20190915175150-20190915201150-00416.warc.gz"}
http://peswiki.com/powerpedia:electric-current
# PesWiki.com ## PowerPedia:Electric current Lasted edited by Andrew Munsey, updated on June 14, 2016 at 9:43 pm. • 112 errors has been found on this page. Administrator will correct this soon. • This page has been imported from the old peswiki website. This message will be removed once updated. Electric current is by definition the flow of electric charge. The SI unit of electric current is the ampere (A), which is equal to a flow of one coulomb of charge per second. #### Introduction In There was an error working with the wiki: Code[13] metal, with no external forces applied, there exists a random motion of There was an error working with the wiki: Code[42]s created by the Thermal energy that the electrons gain from the surrounding medium. When an Atom loses a free Electron, it acquires a net There was an error working with the wiki: Code[43]. The free electron can move amongst these There was an error working with the wiki: Code[44] Ions, while the positive ions can only oscillate about their mean fixed positions. The free electron is therefore the There was an error working with the wiki: Code[45] in a typical solid conductor. Given an imaginary plane through which the wire passes, the number of Electrons moving from one side to the other in any period of time is exactly equal to the number passing in the opposite direction. When a wire is connected across the two terminals of a Direct current There was an error working with the wiki: Code[14], the source places an electric field across the conductor. The moment contact is made, the There was an error working with the wiki: Code[46]s of the conductor will drift toward the There was an error working with the wiki: Code[47] terminal under the influence of this field. For every ampere of current, 1 There was an error working with the wiki: Code[48] of Electric charge (which consists of about 6.242 &times 1018 electrons) drifts every There was an error working with the wiki: Code[49] at the same velocity through the imaginary plane through which the conductor passes. #### Ampere {|align="right" style="width:222px border: solid 2px yellow background: lightyellow hspace:15 vspace:15" | This SI unit is named after There was an error working with the wiki: Code[50]. As for all SI units whose names are derived from the proper name of a person, the first letter of its There was an error working with the wiki: Code[51] is There was an error working with the wiki: Code[52] (A). But when an SI unit is spelled out, it should always be written in There was an error working with the wiki: Code[53] (ampere), unless it begins a sentence or is the name "degree Celsius". &mdash&nbspBased on The International System of Units, section 5.2. |} The ampere (symbol: A) is the SI base unit of Electric current. It is named after There was an error working with the wiki: Code[54], one of the main discoverers of Electromagnetism. ##### Definition The ampere is that constant current which, if maintained in two straight parallel conductors of infinite length, of negligible circular There was an error working with the wiki: Code[55], and placed 1 Metre apart in vacuum, would produce between these conductors a force equal to 2There was an error working with the wiki: Code[1] There was an error working with the wiki: Code[56] per metre of length. Electric current is the time rate of change or displacement of Electric charge. One ampere represents the rate of 1 There was an error working with the wiki: Code[57] of charge per second. :{1 \,A= 1 \,C/s} \, The ampere is defined first (it is a SI base unit, along with the Metre, the There was an error working with the wiki: Code[58], and the There was an error working with the wiki: Code[59]), without reference to the quantity of charge. The unit of charge, the coulomb, is defined to be the amount of charge displaced by a one ampere current in the time of one There was an error working with the wiki: Code[58]. ##### Explanation Because it is a base unit, the definition of the ampere is not tied to any other electrical unit. The definition for the ampere is equivalent to fixing a value of the There was an error working with the wiki: Code[15] of vacuum to ?0 = 4?There was an error working with the wiki: Code[2] H/m. Prior to 1948, the so-called "international ampere" was used, defined in terms of the Electrolysis deposition rate of There was an error working with the wiki: Code[61]. The older unit is equal to 0.999&nbsp85&nbspA. The ampere is most accurately realized using an There was an error working with the wiki: Code[16], the There was an error working with the wiki: Code[17], since the latter two can be tied to physical phenomena that are relatively easy to reproduce, the There was an error working with the wiki: Code[62] and the There was an error working with the wiki: Code[63], respectively. The unit of Electric charge, the There was an error working with the wiki: Code[64], is defined in terms of the ampere: one coulomb is the amount of electric charge (formerly There was an error working with the wiki: Code[65]) carried in a current of one ampere flowing for one There was an error working with the wiki: Code[66]. Current (electricity), then, is the rate at which charge flows through a wire or surface. One ampere of current (I) is equal to a flow of one There was an error working with the wiki: Code[64] of charge (Q) per second of time (t): :{I=Q/t} \, ##### Proposed future definition Since a coulomb is approximately equal to 6.24150948There was an error working with the wiki: Code[3] There was an error working with the wiki: Code[68]s, one ampere is approximately equivalent to 6.24150948There was an error working with the wiki: Code[3] elementary charges, such as Electrons, moving past a boundary in one second. As with other SI base units, there have been proposals to redefine the There was an error working with the wiki: Code[69] in such a way as to define some presently measured There was an error working with the wiki: Code[70] to fixed values. One proposed definition of the kilogram is: : The kilogram is the mass which would be accelerated at precisely 2There was an error working with the wiki: Code[5] m/s2 if subjected to the per metre force between two straight parallel conductors of infinite length, of negligible circular cross section, placed 1 metre apart in vacuum, through which flow a constant current of exactly 6 241 509 479 607 717 888 elementary charges per second. This redefinition of the kilogram has the effect of fixing the There was an error working with the wiki: Code[18] and would result in a functionally equivalent definition for the There was an error working with the wiki: Code[71] as being the sum of exactly 6 241 509 479 607 717 888 elementary charges and the ampere as being the electrical current of exactly 6 241 509 479 607 717 888 elementary charges per second. This is consistent with the current 2002 CODATA value for the elementary charge which is 1.60217653There was an error working with the wiki: Code[6] ± 0.00000014There was an error working with the wiki: Code[6] C. ##### CIPM recommendation There was an error working with the wiki: Code[72] (CIPM) Recommendation 1 (CI-2005): Preparative steps towards new definitions of the There was an error working with the wiki: Code[73], the ampere, the There was an error working with the wiki: Code[19] in terms of fundamental constants The International Committee for Weights and Measures (CIPM), approve in principle the preparation of new definitions and mises en pratique of the kilogram, the ampere and the kelvin so that if the results of experimental measurements over the next few years are indeed acceptable, all having been agreed with the various Consultative Committees and other relevant bodies, the CIPM can prepare proposals to be put to Member States of the Metre Convention in time for possible adoption by the 24th There was an error working with the wiki: Code[20] in 2011 give consideration to the possibility of redefining, at the same time, the mole in terms of a fixed value of the Avogadro constant prepare a Draft Resolution that may be put to the 23rd There was an error working with the wiki: Code[21] in 2007 to alert Member States to these activities #### Current density Current density is a measure of the There was an error working with the wiki: Code[22] whose magnitude is the electric current per cross-sectional area. In There was an error working with the wiki: Code[74], the current density is measured in Amperes per There was an error working with the wiki: Code[75]. Electrical current is a coarse, average quantity that tells what is happening in an entire wire. The current density is an important parameter in There was an error working with the wiki: Code[76] (one of There was an error working with the wiki: Code[77]), which show the direct link between current density and There was an error working with the wiki: Code[78]. Current density is an important consideration in the design of electrical and There was an error working with the wiki: Code[23] systems. Most electrical conductors have a finite, positive There was an error working with the wiki: Code[24] excessive current density may generate a strong enough magnetic field to cause spontaneous loss of the superconductive property. #### Electric charges drift When studying electrical circuits, it is possible that the actual direction of current flow in a specific circuit element is not known at the start. Consequently, we arbitrarily assign each current variable a reference direction. After current values are solved for, some of them might display negative values. Hence, for the negative current variables, the actual current flows in the direction opposite to the reference direction which was originally selected. The mobile charged particles within a conductor move constantly in random directions. In order for a net flow of charge to exist, the particles must also move together with an average drift rate. Electrons are the charge carriers in There was an error working with the wiki: Code[25] are There was an error working with the wiki: Code[26], the waves of electromagnetic energy propagate rapidly through the space between the wires from a source to a distant There was an error working with the wiki: Code[27], even though the electrons in the wires only move back and forth over a tiny distance. Although the velocity of the flowing charges is quite low, the associated electromagnetic energy travels at the speed of light. The ratio of the There was an error working with the wiki: Code[79] through a medium versus the speed of light in a vacuum is called the There was an error working with the wiki: Code[80]. #### Conventional current Conventional current was defined early in the history of electrical science as a flow of positive charge. In solid metals, like wires, the positive charges are immobile, and only the negatively charged There was an error working with the wiki: Code[28] cell may be constructed with salt water (a solution of There was an error working with the wiki: Code[29] are flows of electrons as well as positive and negative ions. In ice and in certain solid electrolytes, flowing There was an error working with the wiki: Code[81]s constitute the electric current. To simplify this situation, the original definition of conventional current still stands. There are also instances where the electrons are the charge that is moving, but where it makes more sense to think of the current as the movement of positive "There was an error working with the wiki: Code[30]" (the spots that should have an electron to make the conductor neutral). This is the case in a p-type There was an error working with the wiki: Code[82]. #### Mathematics The magnitude of an electric current is defined as the There was an error working with the wiki: Code[83] of Electric charge: :I = {dQ \over dt} Formally this is written as: :i(t) = {dq(t) \over dt} or inversely as q(t_0) = \int_{-\infty}^{t_0} i(t)\, dt The amount of charge Q flowing per unit of time t is I, standing for the intensity of the current. The current I in Amperes "flowing" in a wire can be calculated with the following equation: :I = {Q \over t} where :Q \!\ is the Electric charge in There was an error working with the wiki: Code[84]s (ampere seconds) :t \!\ is the There was an error working with the wiki: Code[85] in There was an error working with the wiki: Code[86]s It follows that: :Q=It \!\ and t = {Q \over I} If we want to describe the distribution of the charge flow, we use the concept of the current density: :\vec{J}=nq\vec{v_d}=\rho \vec{v_d} \!\ where :\vec{J} \!\ is the current density vector (SI unit Amperes per There was an error working with the wiki: Code[87]) :n \!\ is the particle density in count per volume (SI unit m-3) :q \!\ is the individual particles' charge (SI unit There was an error working with the wiki: Code[88]s) :\rho = nq \!\ is the There was an error working with the wiki: Code[89] (SI unit There was an error working with the wiki: Code[90]s per There was an error working with the wiki: Code[91]) :\vec{v_d} \!\ is the particles' average There was an error working with the wiki: Code[31] (SI unit Meters per There was an error working with the wiki: Code[92]) The current flowing through a surface S can be calculated by the following relation: :I=\int_S{ \vec{J} \cdot d\vec{S}} &ndash where the current is in fact the integral of the There was an error working with the wiki: Code[93] of the current density vector and the differential surface element d \vec{S}, i.e. the net There was an error working with the wiki: Code[94] of the current density There was an error working with the wiki: Code[95] flowing through the surface S. The speed at which electric charges drift can be calculated from the equation: :I=nAvQ \!\ where :I \!\ is the electric current :n \!\ is number of charged particles per unit volume :A \!\ is the cross-sectional area of the conductor :v \!\ is the drift velocity, and :Q \!\ is the charge on each particle. Electric currents in solid matter are typically very slow flows. ##### Ohm's law There was an error working with the wiki: Code[96] predicts the current in an (ideal) There was an error working with the wiki: Code[97] (or other There was an error working with the wiki: Code[98]) to be applied Voltage divided by Electrical resistance: : I = \frac {V}{R} where :I is the current, measured in Amperes :V is the There was an error working with the wiki: Code[99] measured in Volts :R is the Electrical resistance measured in There was an error working with the wiki: Code[32]s #### Examples Natural examples include There was an error working with the wiki: Code[33] There was an error working with the wiki: Code[34] across long distances and the smaller wires within electrical and electronic equipment. In There was an error working with the wiki: Code[35], and the flow of There was an error working with the wiki: Code[36] within a There was an error working with the wiki: Code[100]. #### Electrical safety The most obvious hazard is electric shock, where a current through part of the body can cause effects from a slight tingle to There was an error working with the wiki: Code[101] or severe There was an error working with the wiki: Code[102]. It is the current that passes that determines the effect, and this depends on the nature of the contact, the condition of the body part, the current path through the body and the voltage of the source. The effect also varies considerably from individual to individual. (For approximate figures see Shock Effects under There was an error working with the wiki: Code[103].) Due to this and the fact that passing current cannot be easily predicted in most practical circumstances, any supply of over 50 volts should be considered a possible source of dangerous electric shock. In particular, note that 110 volts (a minimum voltage at which AC There was an error working with the wiki: Code[37] power is There was an error working with the wiki: Code[38]) can certainly be lethal. Electric arcs, which can occur with supplies of any voltage (for example, a typical There was an error working with the wiki: Code[104] machine has a voltage between the There was an error working with the wiki: Code[105]s of just a few tens of volts), are very hot and emit There was an error working with the wiki: Code[106] (UV) and There was an error working with the wiki: Code[107] (IR). Proximity to an electric arc can therefore cause severe thermal burns, and UV is damaging to unprotected eyes and skin. Accidental electric heating can also be dangerous. An overloaded There was an error working with the wiki: Code[39], There was an error working with the wiki: Code[40], and There was an error working with the wiki: Code[41] are particularly risky because they can deliver a very high current due to their low There was an error working with the wiki: Code[108]. #### Related concepts Alternating current and polyphase currents Direct current and continous currents Electrical conduction There was an error working with the wiki: Code[109] SI There was an error working with the wiki: Code[110] There was an error working with the wiki: Code[111] There was an error working with the wiki: Code[112] #### External articles and references There was an error working with the wiki: Code[9] There was an error working with the wiki: Code[10] There was an error working with the wiki: Code[11] There was an error working with the wiki: Code[12] General Electric current Which direction does electricity really flow? All about circuits - a useful site introducing electricity and electronics, as well as some mathematics involved with circuit calculations. There was an error working with the wiki: Code[1], Wikipedia: The Free Encyclopedia. Wikimedia Foundation. spencerr node46 - A short explanation of the current density Lesson 2: Electric Current Electric Current The NIST Reference on Constants, Units, and Uncertainty A short history of the SI units in electricity @.ampere : Ampère and history of electricity Correspondence, Main books, Movies, 3D VRML - web site create by the CNRS, edited by the Centre de Recherche en Histoire des Sciences et des Techniques : Mme Christine Blondel - CNRS and M. Stéphane Pouyllau - CNRS Books Robert Mullineux Walmsley, The Electric Current, Cassell, 1894 Henry Hutchinson Norris, An Introduction to the Study of Electrical Engineering J. Wiley & Sons, 1909 Franklin Leonard Pope, Modern Practice of the Electric Telegraph. D. Van Nostrand, 1872. Adolf Thomalen, George William Osborn Howe, A Text-book of Electrical Engineering. Arnold, 1907. Newsgroups news://sci.electronics.components Integrated circuits, resistors, capacitors. Activity: Medium. Public - Usenet news://alt.engineering.electrical Discussing electrical engineering. Activity: High. Public - Usenet. news://sci.physics.electromag Electromagnetic theory and applications. Activity: Medium. Public - Usenet.
2017-03-29 23:06:34
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8482034206390381, "perplexity": 1291.2684816355957}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-13/segments/1490218191405.12/warc/CC-MAIN-20170322212951-00014-ip-10-233-31-227.ec2.internal.warc.gz"}
http://clay6.com/qa/34873/number-of-chiral-carbons-in-beta-d-glucose-is
Comment Share Q) # Number of chiral carbons In $\beta -D-(+)$ glucose is $\begin {array} {1 1} (A)\;6 & \quad (B)\;4 \\ (C)\;3 & \quad (D)\;5 \end {array}$ 1. A chiral carbon has four different groups attached. $\alpha$ -D-Glucose has no plane of symmetry, so all the carbon atoms are different.
2019-09-22 20:29:46
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.43316248059272766, "perplexity": 3129.7559264942383}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-39/segments/1568514575674.3/warc/CC-MAIN-20190922201055-20190922223055-00109.warc.gz"}
https://scipost.org/SciPostPhys.5.1.005
## Anomalies for Galilean fields Kristan Jensen SciPost Phys. 5, 005 (2018) · published 20 July 2018 ### Abstract We initiate a systematic study of `t Hooft anomalies in Galilean field theories, focusing on two questions therein. In the first, we consider the non-relativistic theories obtained from a discrete light-cone quantization (DLCQ) of a relativistic theory with flavor or gravitational anomalies. We find that these anomalies survive the DLCQ, becoming mixed flavor/boost or gravitational/boost anomalies. We also classify the pure Weyl anomalies of Schr\"odinger theories, which are Galilean conformal field theories (CFTs) with $z=2$. There are no pure Weyl anomalies in even spacetime dimension, and the lowest-derivative anomalies in odd dimension are in one-to-one correspondence with those of a relativistic CFT in one dimension higher. These results classify many of the anomalies that arise in the field theories dual to string theory on Schr\"odinger spacetimes. ### Ontology / Topics See full Ontology or Topics database. ### Author / Affiliation: mappings to Contributors and Organizations See all Organizations. Funder for the research work leading to this publication
2019-10-23 09:58:13
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3090098202228546, "perplexity": 2510.250402592532}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570987833089.90/warc/CC-MAIN-20191023094558-20191023122058-00220.warc.gz"}
https://gmatclub.com/forum/if-p-and-q-are-positive-p-2-q-2-16-and-p-2-q-2-8-then-q-292929.html
GMAT Question of the Day - Daily to your Mailbox; hard ones only It is currently 21 Apr 2019, 19:18 ### GMAT Club Daily Prep #### Thank you for using the timer - this advanced tool can estimate your performance and suggest more practice questions. We have subscribed you to Daily Prep Questions via email. Customized for You we will pick new questions that match your level based on your Timer History Track every week, we’ll send you an estimated GMAT score based on your performance Practice Pays we will pick new questions that match your level based on your Timer History # If p and q are positive, p^2 + q^2 = 16, and p^2 - q^2 = 8, then q = Author Message TAGS: ### Hide Tags Math Expert Joined: 02 Sep 2009 Posts: 54401 If p and q are positive, p^2 + q^2 = 16, and p^2 - q^2 = 8, then q =  [#permalink] ### Show Tags 09 Apr 2019, 23:01 00:00 Difficulty: 25% (medium) Question Stats: 76% (01:28) correct 24% (01:45) wrong based on 21 sessions ### HideShow timer Statistics If p and q are positive, p^2 + q^2 = 16, and p^2 - q^2 = 8, then q = (A) 2 (B) 4 (C) 8 (D) $$2 \sqrt{2}$$ (E) $$2 \sqrt{6}$$ _________________ Manager Joined: 17 May 2018 Posts: 69 Re: If p and q are positive, p^2 + q^2 = 16, and p^2 - q^2 = 8, then q =  [#permalink] ### Show Tags 09 Apr 2019, 23:18 Subtract eqn2 from eqn2 Get 2q^2=8 Hence q=2 IMO A Posted from my mobile device CEO Joined: 18 Aug 2017 Posts: 3008 Location: India Concentration: Sustainability, Marketing GPA: 4 WE: Marketing (Energy and Utilities) Re: If p and q are positive, p^2 + q^2 = 16, and p^2 - q^2 = 8, then q =  [#permalink] ### Show Tags 10 Apr 2019, 01:57 Bunuel wrote: If p and q are positive, p^2 + q^2 = 16, and p^2 - q^2 = 8, then q = (A) 2 (B) 4 (C) 8 (D) $$2 \sqrt{2}$$ (E) $$2 \sqrt{6}$$ subtract p^2 - q^2 = 8 from p^2 + q^2 = 16 we get q= 2 IMO A _________________ If you liked my solution then please give Kudos. Kudos encourage active discussions. Intern Joined: 11 Jan 2017 Posts: 6 Location: India Re: If p and q are positive, p^2 + q^2 = 16, and p^2 - q^2 = 8, then q =  [#permalink] ### Show Tags 10 Apr 2019, 03:36 p^2+q^2=16 p^2=16-q^2 p^2-q^2=8 16-q^2-q^2=8 q=2 IMO A _________________ Knowledge is Power! Re: If p and q are positive, p^2 + q^2 = 16, and p^2 - q^2 = 8, then q =   [#permalink] 10 Apr 2019, 03:36 Display posts from previous: Sort by
2019-04-22 02:18:43
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.832746148109436, "perplexity": 4935.994121325303}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-18/segments/1555578533774.1/warc/CC-MAIN-20190422015736-20190422041736-00404.warc.gz"}
https://socratic.org/questions/how-do-you-derive-the-ideal-gas-law-from-boyle-and-charles-laws
# How do you derive the Ideal Gas Law from Boyle and Charles’ laws? Feb 29, 2016 Well the original equation is PV = nRT From there, we can make several derivations If any part is constant, we can just isolate it and equal all of these to the other case For example, the R is often constant, so we can set up PV / nT = R = PV / nT Boyle's law assumes that only pressure and volume changes So, we have PV = PV Charles' law assumes that P, R, and n are constant So we have V/T = V/T
2019-11-12 01:11:41
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9461491703987122, "perplexity": 1763.687219644261}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-47/segments/1573496664469.42/warc/CC-MAIN-20191112001515-20191112025515-00322.warc.gz"}
https://www.elitedigitalstudy.com/11167/what-do-you-understand-by-the-term-auto-protolysis-of-water-what-is-its-significance
What do you understand by the term ’auto-protolysis’ of water? What is its significance? Asked by Abhisek | 1 year ago |  82 Solution :- Auto-protolysis (self-ionization) of water is a chemical reaction in which 2 water molecules react to produce a hydroxide ion (OH) and a hydronium ion (H3O+). The reaction involved can be represented as: $$H_{2}O_{(l)}+H_{2}O_{(l)}\leftrightarrow H_{3}O^{+}_{(aq)}+OH^{-}_{(aq)}$$ Auto- protolysis of water indicates its amphoteric nature i.e., its ablity to act as an acid as well as a base. The acid- base reaction can be written as : $$H_{2}O_{(l)}+H_{2}O_{(l)}\leftrightarrow H_{3}O^{+}_{(aq)}+OH^{-}_{(aq)}$$ Answered by Pragya Singh | 1 year ago Related Questions What do you understand by the terms What do you understand by the terms : (i) Hydrogen economy (ii) Hydrogenation (iii) ‘syngas’ (iv) Water-gas shift reaction (v) Fuel cell? How does H2 O2 behave as a bleaching agent? How does H2 O2 behave as a bleaching agent? Do you expect different products in solution when aluminium (III) chloride and potassium chloride treated separately Do you expect different products in solution when aluminium (III) chloride and potassium chloride treated separately with (i) alkaline water (ii) acidified water, and (iii) normal water. Write equations wherever necessary.
2023-03-27 17:32:05
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5395349264144897, "perplexity": 6485.208722377818}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296948673.1/warc/CC-MAIN-20230327154814-20230327184814-00588.warc.gz"}
https://tex.stackexchange.com/questions/132459/bibligraphy-ordering-ignore-entry
# Bibligraphy ordering - ignore entry I am using BibTeX with natbib. My style file orders my bibliography in the order in which articles were cited. Occasionally, however, I want an \cite to be ignored for the purposes of the bibliography order, but fully working otherwise. Minimal working example: \documentclass{article} \usepackage[square,numbers,sort]{natbib} \begin{document} \cite{c} % Want to cite, but ignore for purpose of bibliography order. \cite{a} \cite{b} \cite{c} \bibliographystyle{unsrt} \bibliography{ex} \end{document} with ex.bib: @article{a, author="a"} @article{b, author="b"} @article{c, author="c"} results in: as expected. I want to replace that first citation with a new command, say \ignorecite so that the order is a,b,c, and so that the order in the main document is 3,1,2,3 All you have to do is remove the capability to write a citation to the .aux file. This is achieved by momentarily setting \@fileswfalse: \newcommand{\ignorecite}[1]{{\@fileswfalse\cite{#1}}}% With the entire \cite command still in tact, hyperlinks and formatting functions as usual. \documentclass{article} % Write bibliography items to file \usepackage{filecontents}% http://ctan.org/pkg/filecontents \begin{filecontents*}{ex.bib} @article{a, author="a"} @article{b, author="b"} @article{c, author="c"} \end{filecontents*} \makeatletter \newcommand{\ignorecite}[1]{{\@fileswfalse\cite{#1}}}% \makeatother \usepackage[square,numbers,sort]{natbib}% http://ctan.org/pkg/natbib \usepackage{hyperref}% http://ctan.org/pkg/hyperref \begin{document} \ignorecite{c} % Want to cite, but ignore for purpose of bibliography order. \cite{a} \cite{b} \cite{c} \bibliographystyle{unsrt} \bibliography{ex} \end{document} • That's exactly what I wanted, thanks a lot. Smart solution. Sep 10, 2013 at 13:34 • Is there a similar solution when using biblatex? Oct 31 at 11:36 Similar to Werner's answer, this one by Frank Mittelbach proposes a command that takes the citation call itself: \makeatletter \def\ignorecitefornumbering#1{% \begingroup \@fileswfalse #1% % do \cite comand \endgroup } \makeatother
2022-12-10 01:33:03
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7810721397399902, "perplexity": 5522.871224418361}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-49/segments/1669446711637.64/warc/CC-MAIN-20221210005738-20221210035738-00601.warc.gz"}
https://www.semanticscholar.org/paper/Bifurcation-of-space-periodic-solutions-in-FDEs-Balanov-Wu/da06e0f95835e3231f859e009b9032880068da07
Corpus ID: 73632106 # Bifurcation of space periodic solutions in symmetric reversible FDEs @article{Balanov2016BifurcationOS, title={Bifurcation of space periodic solutions in symmetric reversible FDEs}, author={Zalman Balanov and Haotian Wu}, journal={arXiv: Dynamical Systems}, year={2016} } • Published 30 January 2016 • Mathematics • arXiv: Dynamical Systems In this paper, we propose an equivariant degree based method to study bifurcation of periodic solutions (of constant period) in symmetric networks of reversible FDEs. Such a bifurcation occurs when eigenvalues of linearization move along the imaginary axis (without change of stability of the trivial solution and possibly without $1:k$ resonance). Physical examples motivating considered settings are related to stationary solutions to PDEs with non-local interaction: reversible mixed delay… Expand 3 Citations #### Figures and Tables from this paper Periodic solutions to reversible second order autonomous DDEs in prescribed symmetric nonconvex domains • Mathematics • 2020 The existence and spatio-temporal patterns of $2\pi$-periodic solutions to second order reversible equivariant autonomous systems with commensurate delays are studied using the Brouwer $O(2) \timesExpand Periodic solutions to reversible second order autonomous systems with commensurate delays • Mathematics • 2020 Existence and spatio-temporal patterns of periodic solutions to second order reversible equivariant autonomous systems with commensurate delays are studied using the Brouwer$O(2) \times \GammaExpand Existence and Spatio-Temporal Patterns of Periodic Solutions to Second Order Non-Autonomous Equivariant Delayed Systems • Mathematics • 2020 Existence and spatio-temporal symmetric patterns of periodic solutions to second order reversible equivariant non-autonomous periodic systems with multiple delays are studied under the Hartman-NagumoExpand #### References SHOWING 1-10 OF 25 REFERENCES Reversible Equivariant Hopf Bifurcation • Mathematics • 2005 Abstract.In this paper we study codimension-one Hopf bifurcation from symmetric equilibrium points in reversible equivariant vector fields. Such bifurcations are characterized by a doubly degenerateExpand Hopf Bifurcation in Symmetric Networks of Coupled Oscillators with Hysteresis • Mathematics • 2012 The standard approach to study symmetric Hopf bifurcation phenomenon is based on the usage of the equivariant singularity theory developed by M. Golubitsky et al. In this paper, we present theExpand Hopf bifurcation for equivariant conservative and time-reversible systems We study the bifurcation of small periodic solutions at a non-semi-simple 1:1-resonance in equivariant conservative or equivariant time-reversible systems. By using an equivariant Liapunov-SchmidtExpand Multiple periodic solutions for Γ-symmetric Newtonian systems • Mathematics • 2017 Abstract The existence of periodic solutions in Γ-symmetric Newtonian systems x ¨ = − ∇ f ( x ) can be effectively studied by means of the Γ × O ( 2 ) -equivariant gradient degree with values in theExpand Branches of Periodic Orbits in Reversible Systems In the typical reversible systems which appear in many applications (symmetric) periodic solutions appear in one-parameter families. In this short survey we describe how these branches of periodicExpand On a class of nonlinear Schrödinger equations AbstractThis paper concerns the existence of standing wave solutions of nonlinear Schrödinger equations. Making a standing wave ansatz reduces the problem to that of studying the semilinear ellipticExpand Time-reversal symmetry in dynamical systems: a survey • Mathematics • 1998 Abstract In this paper we survey the topic of time-reversal symmetry in dynamical systems. We begin with a brief discussion of the position of time-reversal symmetry in physics. After definingExpand Reversible Equivariant Linear Systems • Mathematics • 1999 In this paper we classify the structure of linear reversible systems (vector fields) on Rn that are equivariant with respect to a linear representation of a compact Lie group H. We assume theExpand Theory of Degrees with Applications to Bifurcations and Differential Equations • Mathematics • 1997 Elements of Differential Topology. Degree in Finite--Dimensional Spaces. Leray--Schauder Degree for Compact Fields. Nussbaum--Sadovskii Degree for Condensing Fields. Applications to BifurcationExpand Normal forms and unfoldings of linear systems in eigenspaces of (anti)-automorphisms of order two • Mathematics • 2002 Abstract In this article we classify normal forms and unfoldings of linear maps in eigenspaces of (anti)-automorphisms of order two. Our main motivation is provided by applications to linear systemsExpand
2021-12-05 11:42:30
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6209132075309753, "perplexity": 2277.2761898240683}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964363157.32/warc/CC-MAIN-20211205100135-20211205130135-00218.warc.gz"}
https://worldofinterest.wordpress.com/category/finance-2/
Archive | Finance # BOE’s Inflation Report So I had a look at the FT’s coverage of the inflation report, and it brought up one of my pet peeves about the commentary on the crisis, namely, measures of economic slack. The bank of England apparently thinks that there is around 1.5% of slack in the economy. My first pet hate is that “slack” is not adequately defined in this context. Does this mean that Carney thinks that RGDP will develop a new trend line that is only 1.5% higher than the current level and at around 2% a year? Or does this simply mean that at the current moment companies estimate that with their current personnel and facilities they could only increase production by 1.5%. This is important because there are two types of “slack”, there is the measure of “could we do more right now”, and there is the measure “given greater demand could I deploy more employees, capital, and technology to increase supply”. I am pretty sure the BOE’s estimate is about the former, but the macroeconomically important version  is the latter. In other words, the supply side constraints could be soft for a while, if technological improvements since 2008 have enabled greater production, but they just haven’t been deployed due to weak demand. Lets take a quick look at the RGDP (constant prices): Does this look like an economy with only 1.5% slack? Is it not possible that in the medium term we might recover nearly all of our lost productivity? Might we not find that as demand grows aggregate supply just continues to quietly grow so that the BOE’s measure remains stuck at 1.5% while GDP grows 4-5% a year for a number of years? Stolen from an ONS publication, does this look like an economy with only 1.5% slack? Are you telling me that after 7 years of technological progress, the BOE expects slack in GDP to be exhausted before productivity per worker has even reached its 2007 high? Are we really to expect that some (minor) institutional differences will mean that UK workers will not, over the medium term, enjoy the same advances that have allowed US productivity to grow 20% in the same period? The UK is structurally similar to the US, but the structural differences (might) mean that in the UK we get more hiring first, and productivity growth later, and in the US they get more productivity first, and then more hiring. That certainly seems more plausible to me  than that the differences in our institutions mean that the UK will not benefit from the same advances of the US in the long run. In that case, the technology exists already for a 20% rise in productivity in UK workers. The real problem here is that the BOE is giving a short term definition of slack, where it doesn’t assume anything about the future investment and employment and capital improvement programs of enterprises. It just asks about capacity utilisation right now. Which, immediately following a recession, is a sensible enough thing to do. The problem is, that here we have moved into the medium term. There has been technological progress in that time, and the fact that this is waiting in the wings means that we should not regard current utilisation as in anyway a meaningful indicator of “slack”. I predict that productivity per worker will recover its century long trend line of 2% a year growth. In that case, “slack” in the economy is more like 15% than 1.5%. Capital spending, training workers on new technology, the rise of robotics and machine learning will do the heavy lifting over the next 3-5 years, and we can expect strong growth along side benign inflation until we approach this productivity trend line. # An update on Abenomics Via FTAlphaville, I found a graph from this paper: Abenomics is having a huge effect on output compared, which is somewhat masked by demographic decline in the working population. So when will the ECB realise that QE is an effective policy tool which increases output during a depression, and get on with saving the EZ? It didn’t exist. No Seriously. There was no boom in housing construction. The constructing in US housing was exactly what was needed to maintain the housing-population ratio in the face of increased population growth. You cannot have an “unsustainable boom” without oversupply. If you are building exactly the amount that you need, and prices are rising anyway, it is the very definition of a sustainable boom. Its true that housing prices rose and then fell, but they fell exactly the amount that you would expect when there was a peak to trough fall in national income. What does that prove except when people are poorer they will pay less for housing. A change in prices is never the cause of anything, it is always the result of something. The correct way to reason is not “why did an asset crash cause a recession” it is “what caused asset prices to crash”. Aggregate demand is the only story. The world’s central banks let it fall off a cliff in 2008, when they could have prevented it, purely because they were focused on inflation and forgot about nominal income. They assumed that because, historically, inflation and NGDP had been pretty well correlated, that controlling inflation would automatically control aggregate demand at a healthy level. They, in fact, made exactly the mistake that others accuse banks of having made – that they showed too much trust in their theory and didn’t have enough prudence. If the central banks had been watching aggregate demand, this recession could have been a non event. HT to Marcus Nunes for the graph. # Ben Bernanke’s Senate Testimony I always feel that these hearings are usually nothing more than a chance for political leaders to grandstand portray their ignorance. It is usually always easy for a technical expert to come up with an answer which sounds impressive while not really saying a lot, and that most people will be unable to decipher. However, there was one great question, by Klobuchar, which deserved a follow up. (This was well covered by Yglesias): [Mr Bernanke]… what would you have done differently under a single mandate to target inflation and Ben Bernanke hummed and hawed and said, essentially, that he would have done nothing differently. This seems a pretty damning indictment of BB’s policy really, if, with unemployment high and inflation consistently below target, you are not going to take the opportunity to do extra stimulus to lower unemployment, what exactly is the point in a dual mandate? The dual mandate exists because central bankers have known since forever that a little bit of extra demand, which usually creates a little extra inflation, is helpful in lowering unemployment. This is the wisdom of the Philip’s curve, which, for all its flaws, at least underlines the truth that very low inflation is nearly always correlated with high unemployment. Of course, perhaps BB was aiming an under the radar shot at his FOMC hawks: I mean, with the board that he has, perhaps he is unable to force through policy that is expansionary enough for him to feel that he his fulfilling his legal mandate, but he could have chosen to put forwards the hawk’s argument, which would be that excess stimulus would not help unemployment fall any faster. This is a nice argument for this type of hearing, as there is clearly some empirical limit on how fast unemployment can fall – no matter how expansionary, the Fed could not restore full employment in one day. I am completely certain that it could do a hell of a lot better than it is doing, but I accept that falls in unemployment of more than 2% a year are fairly implausible. And if they attempted that they really might cause some inflation. Nevertheless, I cannot help but think that BB’s decision to say nothing was itself an overtly political statement about his feelings on the manner. He is required to go and defend Fed policy, even if he was on the losing side of the FOMC consensus. Just as Mervyn King is forced to talk about how the committee feels there the “costs and risks” of further QE outweigh the benefits, even though we know he voted for more QE. Finally, we can speculate on some nice follow up questions that I might have asked: BB, do you believe that a more expansionary policy would cause unemployment to drop faster, while keeping inflation expectations steady? He would almost be forced to answer yes, and now he would be in a really tight spot, then we could ask BB, given that you believe that expansionary policy would bring down unemployment faster, while maintaining inflation expectations, do you believe that your policy is legal? BB, it is my understanding that the dual mandate exists because of the well known relationship between higher inflation and lower unemployment. Thus, logically, a dual mandate must lead to higher inflation than a single mandate whenever unemployment is significantly above its natural rate. Do you agree? BB, which members of the FOMC do we need to impeach to enable you to legally fulfil your mandate? Wouldn’t that have made great TV! # The Apple Tax Senate Hearings So I did not follow these in particularly great detail, but one thing did catch my ear. Senator McCain accused Apple of pernicious practices which violated the “spirit of the Law”. I really have no time for this argument. If they are interpreting the Law in a manner which the courts uphold as justified, then they are in the clear. In fact, I would say that it is an excuse that politicians use for writing law that is so riddled with contradictions and conflicts as to be essentially garbage. It is literally the job of legislators and lawmakers to write laws which are clear, concise, and unambiguous. It is a basic requirement of competent government, that the powers that be should be capable of writing Law that does what then intend for it to do. Claiming that a company “violated the spirit if not the letter” is just having a tantrum because you aren’t willing to admit your own incompetence. Its childish and disingenuous. # The Ongoing Train-wreck of EU governance: Bundesbank Sues ECB Sometimes it appears that the EU is run by children. The rest of the time it is obviously true. This is one of the latter occurrences. Bundesbank President Jens Weidmann  the anti-hero of the European Depression, has thrown his toys out of the pram. After being out-voted 22-1 on the ECB governing council on instituting the ECB’s bond buying program, he stamped his feet and yelled “I’m telling Mummy”. Mummy’, in this case, being the German Constitutional Court. The OMO program is the one truly successful program the ECB has produced. The mere threat of its existence, announced last June, has calmed sovereign bond markets. Jens’ Weidmann apparently believed that it constitutes direct funding of sovereigns’ and has asked the German Constitutional Court to rule on whether the ECB is violating its own mandate. This all happened last summer, and so why am I bringing it up now? Because the German Court is finally ready to make a ruling. If the court rules against the ECB, what does that even mean? Is a supra-national organisation going to be held hostage to legal rulings of each and every state? What is Draghi to do? What will Merkel do? How will the ECB be held to account? Will it merely ask the Cyprus High court for a dissenting opinion, I am sure they would have no trouble sticking one in Germany’s eye, and then claim that in the absence of a consensus legal opinion they intend to ignore everyone. Will Germany threaten to stop funding the ECB? A comical threat against an institution that holds the printing presses, and in fact, legally owns all of the currency in circulation. Will Germany boycott ECB rate meetings, which I would regard as very good news, as it would almost certainly lead to more policy easing. There is going to be one very bad outcome if this court case comes out against the ECB, the return of massive uncertainty and volatility to peripheral sovereign bonds.  Does anyone want to see the Germany attempt to put another millstone around Spain’s neck? There is nothing good to come out of this, just extra pain for a Eurozone that is continuing to implode in the absence of any serious policy easing. # The Cause of the Crisis Pretty much ever since the crisis of 2008, there are those who have been arguing that the sharp fall in NGDP/demand was the cause of the financial crisis, rather than the other way around. Further evidence has now come to light in both the US and the UK, courtesy of David Beckworth and Britmouse, suggesting that the causes of the crisis go back as far as 2005 at least, and largely vindicating the view that the fall in nominal incomes destroyed banks balance sheets. I strongly urge you to read the Beckworth post. In particular, in the US, expectations of nominal income growth started falling in 2005: By the time Lehman Brothers went under, income growth was already well below trend, with predictable results for default rates. where as in the UK, Nominal income growth has been below trend since 2001, That is a pretty big gap in incomes by 2010. Interpreting why this happened is obviously tough, but I lay the blame on inflation targeting. If you wish to target the price level of outputs, in the face of real cost pressure on your inputs, you must restrict nominal wage growth. Since someones income is someone else’s expenditure, this means that you are restricting NGDP growth. Ever since monetarism was put forward as a theory by Freidmann (and others), they have emphasised that there are two nominal variables which are important, income growth, and inflation. The first because income ultimately determines demand, and the second because nominal wage stickiness makes it hard for employers to adjust wages competitively. The first effort at monetarism, was the targeting of the monetary growth rate, the rational was that both incomes and inflation were strongly correlated to monetary growth. In the US and the UK this turned out to be a transient correlation, and money growth targeting failed (-although it largely succeeded in Germany and Switzerland for nearly thirty years, under a slightly different model, you can get a quick overview from this Mishkin paper). The next logical step was inflation targeting. Why look to target some intermediate variable when you can target them directly. This works just great as long as the two remain strongly correlated. We targeted inflation and got stable income growth for free. :). Of course, this means that had we targeted nominal income/NGDP, we would have got stable inflation for free’. The graphs above demonstrate that this correlation started to break down. This is probably a result of falling world inequality. We outsourced a lot of our labour intensive manufacturing to countries where labour was cheaper. Now income in those countries are rising, and we in the developed world must therefore pay more for the basic raw materials that we use as inputs. Thus, by targeting two percent inflation under rising real costs, we crushed income growth. Over the long term this makes it harder for households to meet their debt repayments. Perhaps the small fall in income growth rate in the UK was stable enough not to contribute materially, but the much sharper fall in the US was probably the cause of the financial crisis. The good news is that central banks can do a lot about this type of nominal problems. By expanding income growth we can put the unemployed back to work. On the other hand, rising wages in the developing world are not going away, so real cost pressures will remain, and higher inflation (in the UK) might be needed in order to keep incomes/demand on a stable trajectory. To my mind, a slightly higher price level is a small cost compared to millions sitting unemployed, so we should do that. This picture also makes sense in of the UK’s productivity puzzle. If raw material costs increase 4%, productivity increases 2% and wage growth is 0%, we get get roughly two percent inflation. However, its not clear that this type of cost inflation is really captured by the productivity statistics. If productivity is measured by output, then it is negative in the above example, since I am producing the same amount of stuff at a higher cost, despite productivity increases. A mixed bag for the Uk then: We must increase nominal wage growth to increase spending, and suffer slightly higher inflation as a result, since the central bank cannot overcome the real cost pressures that are a result of the rising living standards of much of the rest of the world. # Fund Managers and Passive Investing The fund management industry has been getting a lot of bad press recently. Firstly because so many of them fail to beat simple strategies like investing in a FTSE100 index tracker. Of course, in aggregate the fund universe is the stock market, so its clearly impossible for them to beat the stock market in aggregate. Ironically, if all the funds controlled by active fund management where placed in passive funds, it would create exactly the situation needed for the fund management industry to create alpha and thrive. However, its especially hard to understand, because almost every other passive strategy, e.g. buying an equal weighted portfolio instead of a cap weighted portfolio like the FTSE100, will massively outperform a cap weighted index. The Nasdaq-Apple phenomenon illustrates precisely the danger of Cap weighting. At one point Apple comprised nearly 40% of the index (it was later limited to 20% by official decree), which means that you are not getting the benefits of diversification that you would expect in a portfolio, since most of your value is in a handful of positions. The top ten positions of the S&P500 have 20% of the value, you need just 32 to companies to have half the index by value. The bottom fifty companies make up just 1% of the market cap. Is it really worth the trading costs to rebalance 50 positions worth just 0.02% each? Its trivial to see that if every company has identical profiles, the minimum variance portfolio will be equally weighted. We have not even talked about Fee structures. Many funds have total expense ratios exceeding two percent. Truly awful. Investment trusts as a group seem to do better, averaging about 1%. Still, in all this, I came across a surprising graph in the small cap sector: Virtually every small cap fund substantially outperformed the FTSE small cap benchmark over three years. That blue dot over on the left is the benchmark. Risk is vertical and return is horizontal. I have no idea how this can even be true. Perhaps the FT is selecting only the best performers. Perhaps many or most small cap funds are actually only partially invested in small caps? Here is the graph for Flex cap equity. Again the Benchmark seems to be a bizarrely low performer. These graphs were quietly snipped of the FT website, but again, the FTSE all share has indeed managed just 2.5% annualised, so it seems plausible that their flex cap’ index might have done this badly. But how can every fund outperformed so massively? A quick check on my morningstar driven research platform finds 28 small cap funds, but only 9 of them are more than three years old. Does the fund industry routinely close and reopen funds in-order to erase the memory of bad performance? This is really the only possible explanation. No wonder funds look attractive to retail investors when fund managers continually erase fund that under-perform the index, and only those that look good survive. If that is what is happening, the regulators ought to step in, as these graphs are deeply misleading, despite being superficially accurate. However, there is a slight  upside in this for the retail investor. If funds are killed of for relative under performance (and your money returned), then sooner or later you will end up in a fund that is beating the index. Of course, if that out performance is essentially random, then you would be better off indexing and saving your fees, since it is impossible to generate useful work from a random process*. On the other hand, if manager out performance is persistent, and there is reasonable, though not conclusive, evidence that it is, then sooner or later you are likely to end up in a winner. This is interesting enough that I would like to model it. Does anyone have access to fund data including funds that were closed? Ten years would be enough. *Yes, I did just apply the second law of thermodynamics to Finance. It really works too. Try this brain teaser: suppose you wish for a population to have more boys than girls, is there any strategy which can produce this result? Strategies like: I will keep having children till I get a boy. I will continue having children till I have more boys than girls, etc. Strategies involving killing people are not allowed. # Volatility Arbitrage, and the changing markets The Black-Scholes analysis underpins  much of modern financing. Today I am going to take an axe to one of its primary underpinnings. The Black-Scholes analysis depends, among other things, on the idea that volatility can be represented by a single number which is independent of the time over which you mention it. Suppose that we have a random walk made up of randomly sampling a probability distribution, plus a general tendency to move in some direction. I.e. $\frac{dS}{S} = \mu dt + \sigma dX$. Now since dX samples randomly from a distribution with variance $\sigma$, we can expect that it is time dependent. If the variance can be represented by a single parameter, then since $Var(x+y)=Var(x)+Var(y)$, then its clear that for a random walk if we sample over a time that is twice as long, (i.e., contains two samples from the distribution X added together), then we have $\sigma \rightarrow \sqrt{2}\sigma$, and hence we must have $dX\rightarrow\sqrt{2}dt$, thus obtaining Ito’s Lemma by inspection. Now if the market were perfectly efficient, we would know that this must be the case, since if it is not the case, we can obtain an arbitrage opportunity in the following way. (Today we ignore all frictions and transaction costs). Suppose that I have £1,000,000 in cash, and so I decide to take a short position worth £1,000,000 in some index, say the FTSE100 while simultaneously taking a long position in the same index. The only difference is that I am going to rebalance the long position at the open and the close every day to be worth exactly £1,000,000 by using my cash to buy or sell as needed, whereas I will rebalance the short position only every thirty days. The idea is that if volatility is greater on a day to day basis than on a thirty day basis, then by buying high and selling low I can make an incremental profit while, in the main, being hedged against market moves. The open and close data for the FTSE100 is widely available. Ideally I would like to have the tick data so I could look at rebalancing on shorter timescales. Sadly that data appears to be very expensive :(. Anyway, here is the result of the strategy on the FTSE100 since 1984. Notice how the strategy basically fails until 2000, and then is quite successful. I find this graph fascinating. We are looking at the evolution of the market in action. In essence, from 2000, we see the emergence of a term structure, after which the strategy routinely makes money. Note that as we are very close to Beta-neutral, this strategy can be levered up and would need to be, as the absolute returns are not exactly great. However, note that the volatility of the underlying instruments is likely to be much greater than that of the composite index, so the returns would be greater if we applied this strategy individually to each of the constituents. That would mean much more work for me to back test it. 🙂 I suspect that the change in term structure is related to the rise of algorithmic trading. In essence, this strategy is the opposite of trend following. If you trade based on trends, you make the trends shorter and steeper and hence increase volatility. Its plausible that this results in higher volatility on short time scales compared to longer ones. An alternative explanation is that options trading has calmed longer term volatility. Another interesting feature of this term structure applies to options pricing. Since options are priced via Black Scholes based on the volatility, and since measuring volatility on a monthly rather than a daily basis would give you different measures of volatility, then they will also give you different prices. This is already known in finance, where the implied volatility of longer term options is often less than that of shorter term options. Still, its fascinating to me to watch how the evolution of the market and trading ideas has created an inefficiency which did not appear to exist in the past. Probably some Quant based Hedge funds are already trading on this idea, and as it becomes more heavily traded it will disappear. # Valuing Companies: Apple Edition So I am considering a long position in Apple. This post will serve as my investment case, an attempt to apply the mindset of Graham and Dodd. I have a shares Isa, in which I hold a portion of my savings. I try to find undervalued companies and sectors, and aim to hold for a longish time period. My current holdings are BG Group, Man Group, Lamprell, Schroder’s Real Estate Investment Trust, and a hedged Nikki positions through the MSCI Sterling hedged ETF, Cazenove Smaller Companies Fund, and Standard Life Equity Unconstrained. Firstly, I will set out how I think about valuing companies in the abstract. As a shareholder, you live in a gray area of the investing landscape. You have a share in the assets and liabilities of the company, and also a share in current and future profits. I like to think about how a share price reflects a company as a going concern, which means I need to strip out of the market capitalisation those assets and liabilities that are independent of daily operations. Most commonly this involves bond issues. Bonds are more senior than equity, so they get first bite at the profits and also, in the event of bankruptcy, first claim on the assets. Thus, to understand how the market is valuing a company, we need to add its liabilities to its market cap. Similarly, in the unusual event of a company having a very strong net cash position, we should subtract it from the market capitalisation. We can imagine, as it were, that the “market value of a business” = Market Cap + Liabilities – Non productive assets. We should not include such things as factories or offices, or those assets that are necessary for the daily operations of the company. The reasons for this are laid out in Graham and Dodds weighty tome “Security analysis”. When one comes to value a productive asset, its value is usually how much it produces. A company usually only sells such assets when they are no longer profitable, in which case there is no reason to think another company could profitably employ your factory, so its value drops to its land value. As an investor, we cannot know for certain how a company is valuing its productive assets, so in the interests of safety, assume you will get nothing from them that is not fully factored into the profit expectations. Apple produces an interesting test case for this way of thinking, as its strong cash position materially affects its market cap, despite having almost no debt. According to its most recent 10-Q (interim statement for UK readers), it has around $68.7bn in liabilities, and around$158.6bn dollars of non productive assets, mostly in long dated securities and cash. The enterprise value of Apple is therefore its 400bn market cap, less $70bn in net non productive assets, for$330bn. This is the value that should be used for PE ratios. Last years apple earnings were around $41bn consensus estimates suggest it will be around$45bn this year and $50bn the year after. This gives an effective PE ratio of around 8. To put this in perspective, Google is valued at around 24 times earnings, Microsoft and Oracle are measured at around 15 times earnings. (In fairness, these are just the values from the FT website, not the adjusted values I laid out above for apple). This suggests to me that Apple is pretty cheap for a Tech company. Mean reversion is the most powerful force in market dynamics, so I would expect to see this valuation gap resolve itself one way or another. As is often pointed out, the stock market is a forward looking vehicle, so let us ask what this stock price is predicting. To me this suggests that the Stock market is expecting Apple earnings to halve. If they did so then Apple would be fair valued as a tech company at around 15 times earnings. This does not seem plausible to me. It is true that as smart phones have become a commodity margins are likely to come under pressure, and that Google is at a huge premium because the market is very excited about driverless cars and Google glasses. Apple has always been extremely secretive about product development, are we certain that Apple does not have another game changing device up its sleeve? Apple has a fantastic brand, and huge brand loyalty. Smart phones are a fast growing market, and so we could easily see Apple maintaining earnings even with falling margins. Moreover, the corporate culture of excellence that Steve Jobs built is likely to persist. Market Timing So we have set out a value case for Apple. However, market sentiment is a strange beast, and its clear that Apple is in a monster downtrend. Do you fight the trend? Do you catch a falling knife? Well, this is an area in which it helps to look at the big picture. Apple with Support lines and down trend added. So the question is, are we going to break through the support at$420? If it does, we are likely to see another big move lower, down to 380 or so. At the moment the chart looks squeezed between the top of its down trend and the support line. Over the next few days it will break out either higher or lower. The risk of buying now is that it might go  lower. The risk of not buying is that it might break out to the up side by some 20% or so. In such situations my inclination is not to bother too much with the charts. I intend to hold for a long time (two years at least), and so a temporary downwards movement is not such a big deal. Apple is a strongly cash generative company disconnected from its fundamentals (in a good way), so I see a buying opportunity. If your plan is to ride a short move, you have been warned. Disclosure: I will probably initiate a long in Apple later today.
2020-08-11 00:45:57
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 5, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4121352732181549, "perplexity": 1474.974797131145}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439738723.55/warc/CC-MAIN-20200810235513-20200811025513-00531.warc.gz"}
https://brilliant.org/discussions/thread/what-is-the-fault-here/
× # What is the fault here? Find the number of ordered triplets of integers $$x,y,z\ \in\ Z^+$$ such that the L.C.M of $$x,y,z=2^3\times 3^3$$? My approach: Since we are talking about the lease common multiple we can safely assume that all $$x,y,z$$ are of the form $$x=2^a \times 3^d,y=2^b \times 3^e,z=2^c \times 3^f$$ where $$a,b,c,d,e,f$$ are non negative integers.Now we have that $$max(a,b,c)=3,max(d,e,f)=3$$,so at least one of $$(a,b,c)=3$$ and same with $$(d,e,f)$$,now there are three ways of choosing that one,and after that is done the two left can be $$(0,1,2,3$$) hence $$4$$ choices,so the answer should be $$(3 \times 4^2)^2=(48)^2=2304$$,but it isnt,can somebody point out the mistake please. 1 year, 1 month ago Sort by: Siddhartha is correct; your method involves some double-counting. I looked at this case by case: let $$j$$ be the number of elements of $$(a,b,c)$$ that equal $$3$$ and $$k$$ be the number of elements of $$(d,e,f)$$ that equal $$3$$, with both $$j$$ and $$k$$ ranging from $$1$$ to $$3$$. There are then $$3*3 = 9$$ case pairs $$(j,k)$$ to analyze. For example, for $$(j,k) = (1,1)$$ we can choose one of $$(a,b,c)$$ in $$3$$ ways, one of $$(d,e,f)$$ in $$3$$ ways, and the remaining four powers can each be any of $$0,1$$ or $$2.$$ This results in a total of $$3^{6} = 729$$ ordered triples $$(x,y,z)$$. With $$N(j,k)$$ being the number of ordered triples $$(x,y,z)$$ resulting from particular values of $$(j,k)$$, I find that $$N(1,1) = 3^{6}, N(1,2) = N(2,1) = 3^{5}, N(2,2) = 3^{4}, N(1,3) = N(3,1) = 3^{3}, N(2,3) = N(3,2) = 3^{2}, N(3,3) = 1.$$ These values sum to a total of $$1369$$ ordered triples $$(x,y,z)$$. · 1 year, 1 month ago A simpler solution is the square of the number of ways to choose (a,b,c) from (0,1,2,3) minus the number of ways to choose (a,b,c) from (0,1,2). This is $$(4^3 - 3^3)^2 = (37)^2 = 1369$$ · 1 year, 1 month ago well can u please explain the logic behind your answer? i mean how did you get that expression · 1 year, 1 month ago Max(a,b,c) = 3. Therefore each of $$(a,b,c)$$ can take values from $$(0,1,2,3)$$. This is $$4^3$$. Now, we subtract cases where none of $$(a,b,c)$$ are $$3$$. This is equivalent to the cases where values of $$(a,b,c)$$ is from $$0,1,2$$. This is $$3^3$$. Therefore the required value is $$4^3 - 3^3$$.We square this since max(d,e,f) =3 also has $$4^3 - 3^3$$ cases · 1 year, 1 month ago ohh thanks a lot i get it · 1 year, 1 month ago Ohk thanx a lot for helping me! · 1 year, 1 month ago Since we are taking LCM so we can assume x = 2^a 3^d, y = 2^b 3^e, z = 2^c 3^f, where a, b, c, d, e, f, g are non negative integers. As per question max {a, b, c} = 3 and max {d, e, f} = 3. Number of ways in which max {a, b, c} = 3 is 4^3 – 3^3 = 37. Similarly, number of cases for max {d, e, f} = 3 is 37. Hence numbers of required ordered triplets = 37 × 37 = 1369. · 9 months, 3 weeks ago You're double counting many cases. For example, you count the case (3,3,0) two times. · 1 year, 1 month ago
2017-02-25 11:31:16
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7930964827537537, "perplexity": 325.46631037060604}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-09/segments/1487501171758.35/warc/CC-MAIN-20170219104611-00611-ip-10-171-10-108.ec2.internal.warc.gz"}
https://dibyaghosh.com/blog/rl/gcsl.html
# Learning to Reach Goals via Iterated Supervised Learning A demonstration of Goal-Conditioned Supervised Learning (GCSL) # Goal-Conditioned Supervised Learning (GCSL)¶ This notebook provides a brief introduction to goal-conditioned supervised learning (GCSL), as introduced in Dibya Ghosh, Abhishek Gupta, Justin Fu, Ashwin Reddy, Coline Devin, Benjamin Eysenbach, Sergey Levine. Learning To Reach Goals Via Iterated Supervised Learning. In this notebook, we develop a simple instantiation of the GCSL algorithm for solving goal-reaching tasks on a simple 2d navigation task with discrete actions to explain the algorithm. For a codebase with more complete features, please see the Github repository linked below. Let's get started! ## The Goal-Reaching Problem¶ Goal-conditioned supervised learning (GCSL) is an algorithm for solving the problem of goal-reaching in an environment (formally, an MDP). At the beginning of each episode, a goal state $g$ is selected randomly in an environment and shown to the agent. The agent's goal is to reach this goal state by the end of the episode, which terminates after some fixed amount of time has passed. Each step, the agent must choose an action to take, given three pieces of information: 1. The state the agent is currently at 2. The goal the agent is trying to get to 3. The amount of time the agent has left to reach the goal (the horizon) For example, here would be the pseudocode of an agent that takes actions randomly. class RandomAgent: def get_action(self, state, goal, horizon): return np.random.choice(n_actions) # Choose randomly How can we train an agent $\pi: (s, g, h) \mapsto a$ to learn to reach goals in the environment? Imitation Learning: If (somehow) we had access to an optimal policy $\pi^*$, that given any state $s$, goal $g$ and time horizon $h$, could return the optimal action $a^*$, we could just use supervised learning to try to match the actions of the expert (MLE). If actions were discrete, this would correspond to a classification problem, and if actions were continuous, this would correspond to a regression problem. ## Goal-Conditioned Supervised Learning (GCSL)¶ Of course, we don't have access to the optimal policy $\pi^*$ (wouldn't life be simple, then?). What can we do in the absence of this optimal policy? GCSL leverages the insight that even if we don't have access to the optimal policy, we can use the principle of hindsight to turn data that the agent has collected into "optimal" data. The Principle of Hindsight: Suppose the agent has just executed the following episode. It was commanded to reach the goal $g^*$, and ends up seeing the following states and taking the following actions $$s_0, a_0, s_1, a_1, s_2, \dots s_{50}, a_{50}$$ The agent might have done a terrible job at reaching $g^*$. It could be that the final state $s_{50}$ and $g^*$ are on opposite sides of the room. However, we know that the agent was able to successfully reach $s_{50}$, and we can use the data collected in the trajectory to teach the agent how to better get to $s_{50}$. For that matter, the agent was also able to reach $s_{49}, s_{48}, \dots$, and we can use this data to learn how to reach all of these states more accurately. More precisely, consider any two timesteps $t_1$ and $t_2$ ($t_1 < t_2$). We know that after the agent took action $a_{t_1}$ in state $s_{t_1}$, it reached the state $s_{t_2}$ after $t_2-t_1$ timesteps. This means that we can treat $a_{t_1}$ as being optimal behavior at $s_{t_1}$ when trying to reach $s_{t_2}$ in $t_2 - t_1$ timesteps. GCSL: Given the principle of hindsight, we can now introduce GCSL, the algorithm. Each iteration, the agent performs the following three steps: 1. Collects a trajectory with the current agent $\pi_k$ 2. Picks many choices of $t_1,t_2$ to create a dataset $\mathcal{D}$ of optimal data $\mathbf{x} = (s=s_{t_1}, g=s_{t_2}, h=t_2-t_1)$ and $\mathbf{y} = a_{t_1}$. 3. Performs supervised learning (classification or regression depending on discrete actions) to have the policy produce output $\mathbf{y}$ when provided input $\mathbf{x}$ to create a better policy $\pi_{k+1}$ Formally, the policy is performing maximum-likelihood estimation (MLE) via supervised learning. $$\arg\max_\theta \mathbb{E}_{((s,g,h), a) \sim \mathcal{D}}[\log \pi_\theta(a | s=s, g=g, h=h)]$$ ## GCSL in Code¶ Now having provided a general overview of the algorithm, let's see a simple implementation of GCSL with a neural network agent trained in Pytorch, learning to reach goals in a pointmass navigation environment. ## The Environment¶ We use a navigation task in a room to illustrate GCSL. The agent, which gets to observe its $(x,y)$ coordinates, can take one of 5 actions: go up, right, down, left or stay still. The environment defined below is simply a normal Gym env, but with one new function: • env.sample_goal(): To define the goal reaching task, we need a notion of what goals the agent should try to reach. This function samples one such desired goal. In this case, a goal position is chosen randomly in the room. We also implement a helper function for visualizing data from this environment. The agent starts in the blue square, and attempts to reach the orange star. class PointmassEnv(gym.Env): def __init__(self): self.observation_space = gym.spaces.Box(-1, 1, shape=(2,)) self.action_space = gym.spaces.Discrete(5) ... def reset(self): # Resets the agent to the center of the room, and returns state ... def step(self, action): # Given an action in {0, 1, 2, 3, 4} # Take a step in the direction. # Returns (next_state, reward, done, extra_info) ... def sample_goal(self): # Samples a random position in the room ... def plot_trajectory(trajectory): # Plots an episode in the room ... ## The Agent¶ Our agent must choose an action to take, given three pieces of information 1. The state the agent is currently at 2. The goal the agent is trying to get to 3. The amount of time the agent has left to reach the goal (the horizon) Our agent's policy will be determined by a neural network. This agent takes in the state, goal, and horizon; concatenates them all together; and feeds this input into a neural network with two hidden layers of size 100 each. The neural network outputs logits for each action. The agent will sample according to the logits to output the action. class NNAgent(nn.Module): def __init__(self): super().__init__() self.net = nn.Sequential( nn.Linear(5, 100), # Input 2 (state) + 2 (goal) + 1 (horizon) nn.ReLU(), nn.Linear(100, 100), nn.ReLU(), nn.Linear(100, 5), # Output: 5 actions ) def forward(self, state, goal, horizon): horizon = horizon / 50 # Normalize between [0, 1] x = torch.cat([state, goal, horizon], -1) logits = self.net(x) return logits def get_action(self, state, goal, horizon): # Put into PyTorch Notation state_torch, goal_torch, horizon_torch = to_torch(state, goal, horizon) logits_torch = self.forward(state_torch, goal_torch, horizon_torch)[0] probabilities_torch = torch.softmax(logits_torch, -1) probabilities = probabilities_torch.detach().numpy() return np.random.choice(5, p=probabilities) def to_torch(state, goal, horizon): state_torch = torch.tensor(state, dtype=torch.float32)[None] goal_torch = torch.tensor(goal, dtype=torch.float32)[None] horizon_torch = torch.tensor(horizon, dtype=torch.float32)[None, None] return state_torch, goal_torch, horizon_torch ## Acting in the Environment¶ At the beginning of each episode, we choose a desired goal (according to env.sample_goal()). The agent interacts with the environment (in the typical RL fashion) trying to reach this goal by the last timestep of this episode (total 50 timesteps). The function sample_trajectory(env, agent) executes an episode according to the above logic and returns the states seen, the actions the agent took, and what the desired goal was. def sample_trajectory(env, agent, T=50): # Sample a target goal (fixed for episode) desired_goal = env.sample_goal() # Default control loop state = env.reset() states = [] actions = [] for i in range(T): states.append(state) action = agent.get_action(state=state, goal=desired_goal, horizon=np.array(T-i, dtype=float)) actions.append(action) state, _, _, _ = env.step(action) return { 'states': np.array(states), 'actions': np.array(actions), 'desired_goal': desired_goal, } We also write some code to evaluate the performance of the agent. We will measure performance according to how far the agent was from the desired goal at the final state in the episode. • evaluate_agent collects 50 episodes, and prints the median distance to goal for the agent. • visualize_agent visualizes 5 different episodes, each with a different goal, for the agent. def evaluate_agent(env, agent, n=50): distances = [] for _ in range(n): trajectory = sample_trajectory(env, agent) distances.append(np.linalg.norm(trajectory['states'][-1] - trajectory['desired_goal'])) print('Median Distance to Goal: %.3f'% np.median(distances)) print('Min Distance to Goal: %.3f'% np.min(distances)) print('Max Distance to Goal: %.3f'% np.max(distances)) def visualize_agent(env, agent): fig, axes = plt.subplots(1, 5, figsize=(15, 3)) for ax in axes: plot_trajectory(sample_trajectory(env, agent), ax=ax) env = PointmassEnv() agent = NNAgent() evaluate_agent(env, agent) visualize_agent(env, agent) Median Distance to Goal: 0.780 Min Distance to Goal: 0.195 Max Distance to Goal: 1.566 ## Learning a Policy using GCSL (A Code Description)¶ Now that we've established the basics of what an agent is, how it interacts with the environment, and how to evaluate the agent, let's actually train a policy! Suppose that the agent has already collected some trajectories in its buffer. buffer = [] for i in range(10): buffer.append(sample_trajectory(env, agent)) When training the policy, we will sample a trajectory randomly from the dataset trajectory = buffer[np.random.choice(len(buffer))] Choose two time-steps randomly on this path $t_1 < t_2$ (remember, hindsight works for any trajectory and any such pair). t1, t2 = np.random.randint(0, 50, size=2) t1, t2 = min([t1, t2]), max([t1, t2]) If we define • s to be the state at t1 • a to be the action at t1 • g to be the state at t2 • h to be t2 - t1 s = trajectory['states'][t1] a = trajectory['actions'][t1] g = trajectory['states'][t2] h = t2 - t1 Then a is a good action to go from s to g in h timesteps. Because of this, we can treat it as expert data for agent(state=s, goal=g, horizon=h), and try to match it using the classification loss. loss = nn.functional.cross_entropy(agent(s, g, h), a) This leads to the following code-block that fully describes the GCSL algorithm: # Create agent and optimizer agent = NNAgent() learning_rate = 1e-4 optimizer = torch.optim.Adam(agent.parameters(), lr=learning_rate) buffer = [] n_episodes = 201 n_steps_per_episode = 500 for episode in range(n_episodes): # Collect more data and put it in the replay buffer new_trajectory = sample_trajectory(env, agent) buffer.append(new_trajectory) # GCSL optimization for step in range(n_steps_per_episode): # Sample a trajectory and timesteps trajectory = buffer[np.random.choice(len(buffer))] t1, t2 = np.random.randint(0, 50, size=2) t1, t2 = min([t1, t2]), max([t1, t2]) # Create optimal ((s, g, h), a) data s = trajectory['states'][t1] g = trajectory['states'][t2] h = t2 - t1 a = trajectory['actions'][t1] s, g, h = to_torch(s, g, h) a = torch.tensor(a)[None] # Optimize agent(s, g, h) to imitate action a optimizer.zero_grad() loss = nn.functional.cross_entropy(agent(s, g, h), a) loss.backward() optimizer.step() # Print agent performance once in a while if episode % 40 == 0: print('##### Episode %d #####'%episode) evaluate_agent(env, agent) visualize_agent(env, agent) ##### Episode 0 ##### Median Distance to Goal: 0.756 Min Distance to Goal: 0.098 Max Distance to Goal: 1.400 ##### Episode 40 ##### Median Distance to Goal: 0.357 Min Distance to Goal: 0.037 Max Distance to Goal: 1.073 ##### Episode 80 ##### Median Distance to Goal: 0.135 Min Distance to Goal: 0.016 Max Distance to Goal: 0.500 ##### Episode 120 ##### Median Distance to Goal: 0.128 Min Distance to Goal: 0.003 Max Distance to Goal: 0.283 ##### Episode 160 ##### Median Distance to Goal: 0.096 Min Distance to Goal: 0.010 Max Distance to Goal: 0.372 ##### Episode 200 ##### Median Distance to Goal: 0.100 Min Distance to Goal: 0.015 Max Distance to Goal: 0.251 When you run the previous cell, you should see the GCSL agent learning how to reach all the states in the room! And that's it! If you're interested in learning more about how why this simple procedure works, check out our paper. If you're looking to apply the algorithm on larger domains, check out our Github repository Feel free to share!
2020-08-06 08:00:33
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5694572329521179, "perplexity": 4170.8047027718285}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439736883.40/warc/CC-MAIN-20200806061804-20200806091804-00030.warc.gz"}
http://desperatehomeschoolers.com/pages/273302-telecaster-humming-noise
Fenugreek Meaning In Arabic, Salicylic Acid Serum, Crane Bird Nz, Screen Design And Layout In Hci Ppt, Advantages Of Lean Ux, German Potato Salad Bobby Flay, Reasons To Be A Stay-at-home Mom, Nizam College Admissions, … Continue reading →" /> Fenugreek Meaning In Arabic, Salicylic Acid Serum, Crane Bird Nz, Screen Design And Layout In Hci Ppt, Advantages Of Lean Ux, German Potato Salad Bobby Flay, Reasons To Be A Stay-at-home Mom, Nizam College Admissions, … Continue reading →" /> HomeUncategorizedtelecaster humming noise CFLs? So whatever the problem is, it's not a missing or floating ground. If this creates a lot of noise and hum consider new cables, but check and/or fix the guitar ground first. This thing really sounds great. The human body has enough mass to act as a ground when touching ungrounded strings. Suggest you replace it with a better model if a service tech cannot find an obvious problem. The house i'm in right now is very old, i'm not sure what the grounding system in the house is, but many of the wall outlets have no ground (as in only 2 prong outlets). Not even Fender themselves have been able to do that, but it is a decent sound that has all the attributes you would expect. Features How does a guitar normally behave when strings are touched? The guitar is a Fender Telecaster so it's really basic. Change everything in your signal chain ONE THING AT A TIME until you narrow it down to the root cause. i too play a tele and am constantly fighting noise while trying to record on my daw. So I just bought my second electric guitar which is a Fender MIM telecaster. Tons of Color Options. To subscribe to this RSS feed, copy and paste this URL into your RSS reader. i bought a bridge for my tele and the tone is just sublime, its on par with the other boutique broadcaster pickup in term of tone. In all these types of situations you can generally narrow it down by a simple process of elimination. Now I understand telecaster= single coil= hum... not like this though. Hey there! If the buzz stops when you mute your set, the issue is likely audio related. it happens always, but I only really care when i'm recording. Step 1. Humming disappear when I touch the strings. The neck pickup is "covered," in other words, it has a built-in metal shield. Changes when I move I haven't played it in a while, or in the particular house i'm in for that matter but my telecaster is humming to a point where I can't record anything because the hum is so prominent... on a clean channel. Get Rid of the 60Hz Hum on a Guitar. It's soooo awesome and kind of my dream guitar. What else is on in your room? There are two possible circuits used in the Telecaster see these articles for good descriptions: Note that there is usually a ground wire that connects the body of the tone and volume pots. @Sparky256 I understood you to be saying that the strings and pickup coil. This produces a thicker, meatier sound, while still retaining the recognizable characteristics of a Tele. This is not a hum-cancelling combination, but it gives the increased power of a series link. I've tried plugging into a different outlet, turning off all the other electrical devices in the room including the lights, etc. Connecting the coils correctly does NOT ground the strings. The American Performer Telecaster Hum also features Greasebucket(TM) tone circuitry to shape the highs without adding bass, preserving your sound. Simple answer - 60 cycle hum. but the new switch fixed all that. this thought allows me to accept (a certain ammount of) noise and hiss as part of the performance, and maybe even an enhancement. Okay, so my body is an antenna, but why does that cause the guitar to hum, and why don't I cause other electronic devices to hum as I move around the room? This means that when played through an amplifier, there may be some hum or buzz produced by the guitar. The long answer is a lot more complicated and requires more information from you to be able to ignore some of the more unlikely reasons. Then I replace the Amp's power plug. But I can't identify a ground loop here. I play single coil guitars almost exclusively, and I disagree about the need for all that copper shielding and star grounding stuff. I got an american deluxe telecaster two weeks ago, when i bought it, it stated that it had noiseless pick ups, but even if i use it clean or distorted there is this big hum. On reading your question, the replies and comments I do not see where you have definitively identified the source of the hum you are hearing. Schematic here: Try using a short insulated (so that your body doesn't affect it) wire connected to the jack sleeve and probe around the guitar to see if additional grounding helps. Electrical Engineering Stack Exchange is a question and answer site for electronics and electrical engineering professionals, students, and enthusiasts. This guitar does not sound like a 62-64 Telecaster. Flourescent lights, including CFLs? Your always best being plugged into an outlet that doesn't have things like microwaves, bathroom fans, flourescent lights etc on the same circuit. This explanation seems promising but is always presented in a hand-wavy manner. So it's some issue that is at least partially environmental (having to do with either the amplifier or the place) but only affects this guitar, or at least affects it more than it does other single-coil guitars like the Strat. If not, try a different amp, then guitar cable, then guitar. The humming noise stops when the player touches the strings or other metal parts of the guitar. "Loose wire." It is furnished with flush-mount pole pieces for an even string response. I don't know what the hum is like on that, but I feel like it wasn't as bad or else I would have started looking into it earlier. Turn up your Strat (or any guitar equipped with standard single-coil pickups) and you get a certain amount of hum and buzz, especially if playing under fluorescent lights or near anything that creates a strong electromagnetic field. site design / logo © 2020 Stack Exchange Inc; user contributions licensed under cc by-sa. See the update that I added to the question. Admittedly these controls are on expensive Peavy and other amps. You have yet to add enough detailed observations to even begin to guess at the hum source. Another update: Actually the noise isn't a 60Hz hum; it sounds like the 120Hz "angry insect" hum that is often associated with ground loops. The ground for the bridge and strings has gone wrong. Also try other amplifier cables and wiggle the 1/4" plugs at the amp and guitar. How can the pickup coil not be grounded? Telecaster making unbearable Humming noise. This "explanation" is usually followed by advice to "break the loop," try removing/re-installing wires, use a ground-lifting cable or device, or something like that. BTW. When you buy products through links across our site, we may earn an affiliate commission. Start with power.. plug into a different outlet that is known to be on a different circuit breaker. Use a short piece of stranded wire to ground the bridge plate. How to avoid boats on a mainly oceanic world? here is a recording I just took, this is straight into my mac using an mbox 2 no plugins and the laptop was running on battery power. Clean Tele purchased new from Sweetwater within the last 12 months. If the strings are properly grounded there should be no hum whether the strings are touched or not. Asking for help, clarification, or responding to other answers. All of these will cause hum, even LCD monitors and televisions, and in these cases, can cause hum in single coil pickups even when they're off-- you'll have to cut power to 'em. any solutions other than grounding and shielding, i'm not good with wiring, and if it helps i bought the guitar brand new.. (1) is the hum same for all 3 switch positions? The Tele even shows up in metal in the hands of John 5 and Jim Root. CFLs? If this wire is missing or has come loose it must be repaired. I also have a telecaster deluxe, although it's getting fixed right now at a friends house because it constantly slips tuning. The Fender Telecaster is a legendary instrument, but it can become frustratingly noisy at times, leaving you frantically searching for a way to quiet it down. The answer is "grounding and shielding" and advises checking wires, changing components, etc. There are two major ways the hum gets in to these circuits: conducted and radiated. The problem is the humming. If they both weren't attached to the right places, the guitar wouldn't work at all. The hum is worse with the bridge pickup, which is interesting, because of course that's the pickup that was shielded by the bridge cover that Fender originally fitted to the guitar but that everyone took off and used as an ashtray. If I play with it in a band practice with a lot of gain (i know it's not the "best" idea playing with tele) it hums quite a lot. Also, things in other rooms can cause it. i find it helpful to remember that i am trying to use pre-wwII technology to interface with advanced digital-age technology. Quote: Originally Posted by sonic dogg The fact is the wiring of guitars is lacking in so many ways. The cable sleeve is in turn connected to the metal amplifier chassis and ultimately to the mains ground. Features How to Get Rid of an Unwanted Guitar Buzzing Noise. If your guitar has this issue the ground will need to be re-soldered. That tunes out the hum fields via a trimpot and gives my Tele's an 80 db signal to noise ratio, a very good spec. None of what you say is wrong in any way. CFLs? It's shown also in available articles of Telecaster wiring. I have seen and fixed this problem enough to say it is common with certain types of guitars and old hand-me-down guitars. : Ok guys, here is a simple way to get rid of the annoying 60Hz hum that many guitar player will experience at one time or another. What is the application of rev in real life? Doesn't your response contradict the one from @Sparky256? Seriously, your idea to try another amp points to your amp as being defective. If it comes loose the guitar will have no signal ground! I think your issue is ground, but shielding ain't evil at all. (2) have you tried it outdoors? But the strings are still connected to the other metal parts of the guitar except the signal circuit and you can be grounded via some other route, for ex you have leather shoes and you stand on concrete floor. Still, I disagree with the poster I quoted only in that the shielding ain't going to hurt the tone. Often mains overhead power lines. What's the best way for EU citizens to enter the UK if they're worried they might be refused entry at the UK border? I have a Les Paul that sat in the case for several years that hummed and howled like a banshee when I plugged it in after a long hiatus. "Your body is an antenna/capacitor plate." In that case the strings aren't even part of the circuit so why would touching them do anything at all? Create a username and password below and an account will be created and your post entered. Or should I try playing somewhere else? EDIT: Based on the diagram you provided the white wire from the Jack is signal ground, but the bridge plate part#21 should be connected to this white wire to ground the strings. The STL-1 vintage pickup has the most iconic sound on a Telecaster. If you were to wire your house with the techniques employed for all these years you would have burnt it to the ground by now. Do you use dimmer switches on the lights in your house? The person providing this answer recommends checking solder connections etc. The hum is making a similar sound to that of a guitar cable that is not plugged in, the weird thing is when my cable is unplugged, and I move it around the room, point it at different areas, the hum changes. Does your amp or guitar have a toggle switch for phase or ground options? DeepMind just announced a breakthrough in protein folding, what are the consequences? Ground wire? I know it's "mains current" or something like that. It has been specifically made for a Telecaster. Making statements based on opinion; back them up with references or personal experience. Well, this guitar features a humbucker in the neck position, just the way Keith Richards likes it. @WillisBlackburn My answer clearly states that the strings and pickup coil MUST be connected to signal ground in the guitar or the result could be severe hum and noise. By clicking “Post Your Answer”, you agree to our terms of service, privacy policy and cookie policy. There is a small blemish on one fret that I tried my best to picture. Neon lights? Now I understand telecaster= single coil= hum... not like this though. The pickup only has two wires. Comment on your update: I don't know if the stings are connected to the signal ground or not. In this case, check all of the TV's external audio connections, including composite and coaxial cables. This suggests that you are in a country outside of the EU. Single coils are noisy, hence the name ‘humbucker’ because the dual coil bucks the hum. I would like to understand what is actually happening. Not all guitars have proper grounds installed. The humming noise sounds like 120Hz hum (see update below). High thin buzzing from Fender Telecaster? Why is a third body needed in the recombination of two hydrogen atoms? I'm skeptical that it's a wiring issue with my guitar since touching different parts doesn't affect it, the only way I can seem to change it is by pointing the pickups at different things, which leads me to believe my house is getting some crazy electro magnetic interference. MathJax reference. Fender Vintage Noiseless Tele pickups produce the brilliant single-coil clarity, definition, and twang of a vintage 1960s Tele without the hum. If you find this sound irritating, as many do, you might consider a guitar with humbucking pickups. I checked the grounding and confirmed that there is continuity all the way from the strings to the cable sleeve to the amp chassis and to the ground in the wall power socket. After you're done doing all of that shield it. "Not enough shielding." also, check the "dumb guy" stuff - make sure you don't have anything metallic in your pockets, don't play near any kind of rf, radio or tv receptors. Any color of stranded small gauge wire will do. I have a Fender Telecaster that hums loudly--much more so than other similar guitars. Another type of pickup uses a separat… Update to this question, 21-Sep-2019: I was able to do more investigation of this issue. also, check for simple things - turn off cellphones, tv's, anything electrical that doesn't need to be on. Can you explain why the hum goes away when I touch the strings? Name: Rudy Wijaya Rating: 5 Broadcaster pickup but with minimum amount of 60 cycle hum what a great broadcaster pickup! Noise gates can help keep the noise under control when you are not playing, but when you hit a note, and the gate opens, the noise … i think it was all wrong. I wouldn't be suprised if the 3 prong outlets weren't actually connected to a gound. This explanation suggests the there is some potential being generated in the player's body that is being transferred to the pickup and that touching the strings grounds the player. I had an old bass guitar with similar problems but never really got to the bottom of it. It's possible that the strings just in your Telecaster are NOT connected to the signal ground, so it differs from my Telecaster. I have tried using it on a tube amp with and w/o a ground lift, on my laptop using my interface with the computer plugged in/ on batter power and the hum never changes. I'll check with a multimeter and see what's connected to what. I tried touching all of the metal contacts on the guitar to see if something wasn't grounded properly, but it all stayed constant. I have a Fender Telecaster that hums loudly--much more so than other similar guitars. This can involve putting shims under the bridge or neck pickups. You said you haven't played in a while, maybe you just need to clean the jack. If you’re hearing noise, buzz or hum that’s as loud or louder than your guitar, you may have a bad ground inside your guitar. I've had to position guitarists when in the control room because the pickups were catching its own signal through the monitors. Why do Arabic names still have their meanings? Just a friendly reminder that political discussion, (including "offhand" and 'sideways' commenting) is. Here's the kicker whenever I move around, point the guitar up down rotate, the hum changes significantly, If i play with the guitar pointed to the ceiling I barely get a hum. If yes, then any difference? My solution? How do EMH proponents explain Black Monday (1987)? No noise, original pickups and sound = … Why is training regarding the loss of RAIM given so much more emphasis than training regarding the loss of SBAS? I have been given a fantastic present - an American Deluxe Fender Telecaster along with a Line 6 Spider III 30W amp. What is causing the hum? Solution: Repair Ground. QUESTION. It's just a guitar plugged into an amplifier. Some amps do have a ground polarity switch or ground phase control.
2021-01-15 17:36:27
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.23487812280654907, "perplexity": 2036.1556675613704}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610703495936.3/warc/CC-MAIN-20210115164417-20210115194417-00423.warc.gz"}
https://github.com/wtadler/confidence/
Use experiments, modeling, and simulation to determine if perceptual confidence emerges from Bayesian or heuristic computations. 🧠 Fetching latest commit… Cannot retrieve the latest commit at this time. Type Name Latest commit message Commit time Failed to load latest commit information. calibration helper_functions human_data model_fits neuralnet neuralnet2 signatures .gitignore categorical_decision.m optimize_fcn.m run_categorical_decision.m top_ten.mat Code (papers 1–3) and behavioral data (papers 1–2) MCMC model fits to behavioral data (papers 1–2) # Are human confidence reports Bayesian? This repository accompanies the following three peer-reviewed papers on the topic of whether human confidence reports are Bayesian, by Will Adler, Rachel Denison, Marisa Carrasco, and Wei Ji Ma: ##### 1. Confidence when uncertainty is due to stimulus reliability William T. Adler, Wei Ji Ma. (2018). Comparing Bayesian and non-Bayesian accounts of human confidence reports [pdf]. PLoS Computational Biology. ##### 2. Confidence when uncertainty is due to inattention Rachel N. Denison*, William T. Adler*, Marisa Carrasco, Wei Ji Ma. (2018). Humans incorporate attention-dependent uncertainty into perceptual decisions and confidence [pdf]. Proceedings of the National Academies of Sciences. ##### 3. Theoretical exploration of Bayesian confidence signatures. William T. Adler, Wei Ji Ma. (2018). Limitations of proposed signatures of Bayesian confidence [pdf]. Neural Computation. Each of the above papers is associated with a chapter in Will Adler's 2018 dissertation in neural science at NYU. # Usage ## Papers 1 and 2 The below instructions apply to Papers 1 and 2. The code for Paper 3 is relatively straightforward and you can find a different README with the corresponding code. ### Setup Clone or download this repository. In MATLAB, cd into this repository. Use addpath(genpath('helper_functions')) to add the necessary functions to your path. ### Plot some data Use show_data_or_fits to plot the human data. Try this to plot data from Paper 1: axis.col = 'slice'; axis.row = 'depvar'; 'axis', axis, ... 'depvars', {'resp', 'Chat', 'tf'}, ... 'slices', {'c_s', 'c', 'c_C'}); To get two figures. The second one, for Task B, should look like this: The parameters in axis indicate that: • The columns of the resulting figures should show different slices of the data. In this case, the data in the first column slices the data by reliability c and orientation s (slice c_s); the second column slices the data by just reliability c (slice c); and the third column slices the data by reliability c and true category C (slice c_C). • The rows of the resulting figures should show different dependent variables on the y-axis. In this case, those variables are mean button press (resp), proportional response "category 1" (Chat, as in C-hat, used in the papers), and proportion correct (tf, as in true-false). • You will get a figure like this for each task. In this case, a dataset was chosen that includes two tasks, and both tasks A and B were specified as arguments, so you will get two figures. Options for the axis parameters are 'none', 'slice', 'depvar', 'task', 'subject', and 'model'. 'subject' will give you individual subject data rather than grouped, as shown above. 'model' will show you fits to different models (more on that below). Options for depvars are: • resp Mean button press (choice and confidence), between 1 and 8 • Chat Proportion response "category 1" • tf Proportion correct • g Mean confidence, between 1 and 4 • rt Reaction time (s) Options for slices are 's', 'Chat', 'g', 'resp', 'rt', 'C_s', 'c', 'c_s', 'c_C', 'c_Chat', 'c_g', and 'c_resp', which are just various combinations of the variables described above. To plot data from Paper 2, try running this code: axis.col = 'slice'; axis.row = 'depvar'; axis.fig = 'none'; 'axis', axis, ... 'depvars', {'resp', 'Chat', 'tf'}, ... 'slices', {'c_s', 'c', 'c_C'}); to get something like this: ### Explore model fits If you want to explore model fits, try going to this repository and downloading the .mat files into the model_fits folder. Be warned that these are 11 files averaging about 600MB. They're big because they include the thinned chains of samples generated when we fit the models with MCMC. Once downloaded, run load_model_fits.m to load in all the files and organize the models as they are used in Paper 1. Now you will have several loaded structs containing all you need to know about model parameters (e.g., lower and upper bounds) and fits (e.g., MCMC samples and scores to each subject). For instance, the attention.modelmaster struct will contain all the models used in Paper 2. Try loading and printing out the names of the models: load_model_fits models = attention.modelmaster; rename_models(models); >>> 1: Bayes_{Strong}-{\itd}N + non-param. \sigma, B >>> 2: Bayes_{Weak}-{\itd}N + non-param. \sigma, B >>> 3: Orientation Estimation + non-param. \sigma, B >>> 4: Linear Neural + non-param. \sigma, B >>> 5: Lin + non-param. \sigma, B >>> 6: Quad + non-param. \sigma, B >>> 7: Fixed + non-param. \sigma, B >>> 8: Bayes-{\itd}N + non-param. \sigma, B, choice only >>> 9: Orientation Estimation + non-param. \sigma, B, choice only >>> 10: Linear Neural + non-param. \sigma, B, choice only >>> 11: Lin + non-param. \sigma, B, choice only >>> 12: Quad + non-param. \sigma, B, choice only >>> 13: Fixed + non-param. \sigma, B, choice only If you're interested in the Quad confidence model (model #6), for instance, here's some of the information you could pull out of the struct: • models(6).parameter_names: names of the model parameters • models(6).lb: parameter lower bound values • models(6).ub: parameter upper bound values • samples = vertcat(models(6).extracted(1).p{:}): MCMC samples (nSamples x nParams) for the model, subject 1. (Each cell in models(12).extracted(1).p is an MCMC chain) With the models loaded, you can also use show_data_or_fits to plot model fits and scores. Try running this code: axis.col = 'model'; axis.row = 'depvar'; axis.fig = 'none'; 'axis', axis, ... 'depvars', {'resp', 'tf'}, ... 'slices', {'c_s', 'c', 'c_C'}, ... 'models', models([1:2 5:7]), ... 'MCM', 'loopsis'); to get a plot that includes model fits as well as PSIS-LOO score comparisons for five models: ### Understanding the computations behind each model There are two functions that you might want to look through if you are interested in how the computations described in the paper are implemented in the code. One function is trial_generator.m, which shows how fake data is generated for each model. The other one is nloglik_fcn.m which is used for computing the (negative) log-likelihood of each model. ### Run the psychophysical experiment To run the experiment, you'll need to install Psychtoolbox. Then run_categorical_decision.m is used to set the parameters of the experiment and run it. ## Paper 3 All of the code necessary to produce the simulations in Paper 3, as well as a README specific to Paper 3, can be found in the signatures folder. Paper 3 involves no human data. You can’t perform that action at this time.
2019-03-20 16:26:08
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4153349697589874, "perplexity": 9060.070597136075}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-13/segments/1552912202433.77/warc/CC-MAIN-20190320150106-20190320172106-00438.warc.gz"}
https://www.mathimatikoi.org/forum/viewtopic.php?f=28&t=938
It is currently Fri Feb 22, 2019 7:30 pm All times are UTC [ DST ] Page 1 of 1 [ 2 posts ] Print view Previous topic | Next topic Author Message Post subject: The dual space of $X$ giving information about $X$Posted: Thu Jul 14, 2016 10:48 am Team Member Joined: Tue Nov 10, 2015 8:25 pm Posts: 314 Let $\displaystyle X$ be a normed space and let $\displaystyle X^{*} = B(X,\mathbb{R})$ be its dual space. Show that 1. For all $\displaystyle x,y \in X$ such that $\displaystyle x \neq y$ there exists $\displaystyle f \in X^{*}$ such that $\displaystyle f(x) \neq f(y)$. $\\$ 2. If $\displaystyle Y$ is a proper closed subspace of $\displaystyle X$ and if $\displaystyle x_{0} \in X \smallsetminus Y$, then there exists $\displaystyle f \in X^{*}$ with $\displaystyle ||f||=1$ such that $\displaystyle f(y)=0,\forall y \in Y$ and $\displaystyle f(x_{0}) = d(x_{0},Y)$. $\\$ 3. If $\displaystyle X^{*}$ is seperable, then so is $\displaystyle X$. Is the converse true? Top Posted: Thu Jul 14, 2016 10:48 am Team Member Joined: Mon Nov 09, 2015 1:52 pm Posts: 426 Hi Nickos. 1. Let $\displaystyle{x\,,y\in X}$ such that $\displaystyle{x\neq y}$. Let $\displaystyle{u=x-y\neq \overline{0}}$. Consider the subspace $\displaystyle{Y=<u>=\left\{k\,u: k\in\mathbb{R}\right\}}$ of $\displaystyle{\left(X,||\cdot||\right)}$ . Obviously, $\displaystyle{\dim_{\mathbb{R}}Y=1}$ . We define $\displaystyle{g:Y\longrightarrow \mathbb{R}}$ by $\displaystyle{g(k\,u)=k\,||u||}$ . Let $\displaystyle{y_1\,,y_2\in Y}$ and $\displaystyle{a\in\mathbb{R}}$. So, $\displaystyle{y_1=k_1\,u\,\,,y_2=k_2\,u}$ for some $\displaystyle{k_1\,,k_2\in\mathbb{R}}$ and then $\displaystyle{g(y_1+y_2)=g\,((k_1+k_2)\,u)=\left(k_1+k_2\right)\,||u||=k_1\,||u||+k_2\,||u||=g(y_1)+g(y_2)}$ $\displaystyle{g(a\,y_1)=g\,(a\,(k_1\,u))=g\,((a\,k_1)\,u)=(a\,k_1)\,||u||=a\,(k_1\,||u||)=a\,g(y_1)}$ which means that $\displaystyle{g}$ is a linear map. Also, if $\displaystyle{y=k\,u\in Y\,\,,k\in\mathbb{R}}$, then : $\displaystyle{\left|g(y)\right|=\left|g\,(k\,u)\right|=\left|k\,||u||\right|=\left|k\right|\,||u||=||k\,u||=||y||}$ . According to $\displaystyle{\rm{Hahn-Banach}}$ - theorem, there is $\displaystyle{f:X\longrightarrow \mathbb{R}}$ with $\displaystyle{f\in X^{\star}}$, $\displaystyle{\left|f(x)\right|\leq ||x||\,,\forall\,x\in X}$ and $\displaystyle{f(y)=g(y)\,,\forall\,y\in Y}$ . $\displaystyle{f(u)=g(u)=g\,(1\cdot u)=1\cdot ||u||=||u||\neq 0}$ so : $\displaystyle{f(x-y)\neq 0\iff f(x)-f(y)\neq 0\iff f(x)\neq f(y)}$ . 2.We observe that $\displaystyle{x_0\neq \overline{0}}$ cause $\displaystyle{x_0\notin Y}$ . So, we define $\displaystyle{Z=<Y\cup\left\{x_0\right\}>}$ , $\displaystyle{g:Z\longrightarrow \mathbb{R}\,,g(y+k\,x_0)=k\,d\,(x_0,Y)}$ and we continue as in 1. 3.The set $\displaystyle{S=\left\{f\in X^{\star}: ||f||=1\right\}}$ is a seperable subset of $\displaystyle{\left(X^{\star},||\cdot||\right)}$ cause $\displaystyle{\left(X^{\star},||\cdot||\right)}$ is seperable (left as an exercise) . There exists dense subset $\displaystyle{D=\left\{f_{n} :n\in\mathbb{N}\right\}}$ of $\displaystyle{S}$ , which means that $\displaystyle{\left(\forall\,s\in S\right)\,\left(\forall\,\epsilon>0\right)\,\left(\exists\,d\in D\right): ||d-s||<\epsilon}$ . Exercise: Prove that : $\displaystyle{\left(\forall\,n\in\mathbb{N}\right)\,\left(\exists\,x_{n}\in\left\{x: ||x||\leq 1\right\}\right) : f_{n}(x_n)>\dfrac{1}{2}}$ . Setting $\displaystyle{Y=<\left\{x_{n}: n\in\mathbb{N}\right\}>}$ , we have that $\displaystyle{Y\simeq \mathbb{N}}$ and $\displaystyle{\overline{Y}=X}$ (why ?) , which means that $\displaystyle{\left(X,||\cdot||\right)}$ is seperable. The converse is not true cause $\displaystyle{\mathbb{l_{1}}^{\star}=\mathbb{l}_{\infty}}$ and $\displaystyle{\mathbb{l}_{\infty}}$ is not seperable. Top Display posts from previous: All posts1 day7 days2 weeks1 month3 months6 months1 year Sort by AuthorPost timeSubject AscendingDescending Page 1 of 1 [ 2 posts ] All times are UTC [ DST ] #### Mathimatikoi Online Users browsing this forum: No registered users and 1 guest You cannot post new topics in this forumYou cannot reply to topics in this forumYou cannot edit your posts in this forumYou cannot delete your posts in this forumYou cannot post attachments in this forum Search for: Jump to:  Select a forum ------------------ Algebra    Linear Algebra    Algebraic Structures    Homological Algebra Analysis    Real Analysis    Complex Analysis    Calculus    Multivariate Calculus    Functional Analysis    Measure and Integration Theory Geometry    Euclidean Geometry    Analytic Geometry    Projective Geometry, Solid Geometry    Differential Geometry Topology    General Topology    Algebraic Topology Category theory Algebraic Geometry Number theory Differential Equations    ODE    PDE Probability & Statistics Combinatorics General Mathematics Foundation Competitions Archives LaTeX    LaTeX & Mathjax    LaTeX code testings Meta
2019-02-22 18:30:16
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 2, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7042200565338135, "perplexity": 4699.582453143937}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-09/segments/1550247522457.72/warc/CC-MAIN-20190222180107-20190222202107-00572.warc.gz"}
https://www.nature.com/articles/s41612-021-00188-5?error=cookies_not_supported&code=a9ef7be5-fdcf-47d7-a673-801b84094c2c
## Introduction Over the 1980–2012 period, the eastern tropical Pacific sea surface temperature (SST) is characterized by a cooling trend that was one of the main causes of the global surface warming slowdown observed during 1998–20121,2,3. This regional cooling contrasts with a direct radiatively forced response expected from the increase in anthropogenic greenhouse gases4 and it is associated with an intensification of the western tropical Pacific easterlies, reflecting changes in the Walker Circulation5,6,7. Such changes have been partly attributed to variations in the tropical Atlantic SST through atmospheric teleconnections8,9. During the same 1980–2012 period, the tropical Atlantic SSTs continued warming, likely due to a combination of anthropogenic-related radiative forcing and internal climate variability10,11. In particular, the leading mode of decadal variability of the North Atlantic SST—namely the Atlantic multidecadal variability (AMV12,13; Fig. 1a, b)—shifted from a cold to a warm phase around 1995–1996, exaggerating the North Atlantic warming trend induced by anthropogenic greenhouse gases14,15. Over the longer 1920–2014 period, warm AMV conditions were also associated with cold SST anomalies in the central and eastern tropical Pacific (Fig. 1c), supporting the existence of a consistent link between the AMV and the tropical Pacific climate16,17. Yet, these observed Pacific changes cannot be unequivocally attributed to the AMV due to the presence of external forcing and internally driven variability outside of the North Atlantic, as well as because of the limited historical record with respect to the timescales considered and observational uncertainties. Coupled global climate model (CGCM) simulations offer the possibility to tackle these limitations. Using a hierarchy of numerical models, Li et al.9 demonstrated that the tropical Pacific response to the Atlantic forcing can be decomposed into two phases: Phase-1 an initial Atlantic forcing through diabatic heating and Phase-2 an Indo-Pacific Walker Circulation feedback (cf. Fig. 3 in Li et al.9). In Phase-1, the warm tropical Atlantic SST anomalies in summer (hereafter seasons are relative to the Northern Hemisphere) intensify deep convection and lead to upper tropospheric mass divergence over the tropical Atlantic. This divergence is compensated by upper tropospheric mass convergence and descent over the Central tropical Pacific, which intensifies the surface Trade winds over western tropical Pacific8,18. In Phase-2, the so-called Indo-Pacific feedback reinforces the Trade winds, piling up warm water in the Pacific Warm Pool, where atmospheric deep convection increases. This results in an upper tropospheric mass divergence over the warm pool that enhances Central tropical Pacific descent acting as positive feedback on the anomalies generated by the Atlantic forcing in Phase 19,19. Following El Niño Southern Oscillation (ENSO) dynamics, an increase in summer easterlies in the western tropical Pacific eventually favors colder conditions than normal in the eastern and central Pacific during the following winter20. Given the global importance of the tropical Pacific variability and the predictability arising from the North Atlantic at decadal timescales21,22, this Atlantic–Pacific teleconnection is a potential source of seasonal to decadal climate predictability that needs to be further assessed in models. However, the robustness and the strength of this connection remain unknown and need to be quantified. Here, we present a multi-model assessment of this Atlantic–Pacific connection using 21 ensemble simulations from 13 CGCMs (Supplementary Tables 1 and 2) that largely comply with the CMIP6/DCPP-C protocol23. Following this protocol, the same observed AMV SST anomalies (Fig. 1a) are imposed in the North Atlantic of each CGCM to investigate the worldwide teleconnections associated with the observed AMV (see “Methods”). We note that in those idealized AMV simulations, extra heat is added to (or removed from) the climate system to maintain a stationary AMV signal in the North Atlantic for 10 years. This artificial heat prevents a realistic simulation of the relationship between AMV and the global mean surface temperature. ## Results ### Uncertainty in the Pacific response to AMV forcing We start by discussing the multi-model mean (MMM; cf. “Methods”) winter response of the AMV experiments. Associated with the imposed 0.2 °C tropical North Atlantic warming, the MMM shows a 0.05 °C cooling in the tropical South Atlantic and a 0.1 °C cooling in the central equatorial Pacific (Fig. 2a). The latter extends eastward and poleward in both hemispheres, contrasting with warm anomalies in the western part of the subtropical Pacific basins. In the Indian Ocean, the MMM shows a broad warming response with maximum anomalies localized west of India. The summertime SST anomalies are similar to the winter ones but of weaker amplitude over the central equatorial Pacific (Fig. 2b). Overall, the MMM shows good agreement with observations over the whole tropical Atlantic (even south of the Equator where models are not constrained) as well as North of 10°S in other tropical regions (Fig. 1c). This similarity supports the important driving role of the AMV in the observed changes over the Pacific during the historical period8,9,17,19,24,25,26,27. In addition, the negative response of the tropical Pacific SST to the imposed North Atlantic warming in the AMV experiments implies a dynamical adjustment of the Pacific. We now investigate the tropical Pacific response as simulated by each individual model using as a proxy the NIÑO3.4 SST index (cf. indices definition in “Methods” and Fig. 2a). In winter, all experiments simulate La Niña-like cooling in response to an AMV warming except the EC-Earth3P_1Sig experiment that shows weak NIÑO3.4 warming of +0.01 °C (Fig. 3a; see also Supplementary Fig. 8). Though models mostly agree on the sign of the tropical Pacific response, the magnitude of their response varies by an order of magnitude, from 0.01 °C to −0.23 °C, with a MMM of −0.12 °C for a similar ~0.2 °C tropical North Atlantic warming. This large inter-model spread in response to AMV forcing highlights considerable uncertainties in our ability to predict the climate at seasonal to decadal timescales28,29. ### Origins of the inter-model spread Different tropical Pacific responses among models in winter can be explained by intrinsic model differences in simulating Pacific climate dynamics such as the ones linked to ENSO30. Yet, it is known that ENSO is strongly influenced by tropical Pacific conditions in the previous summer31,32,33,34. In particular, tropical Pacific heat content anomalies and their driving surface winds are known to be predictors of ENSO several months ahead35,36. Therefore, different tropospheric responses to the Atlantic SST forcing during summer can also explain model differences in winter20,37. Here, we find that the winter NIÑO3.4 inter-model spread is mainly associated with the inter-model spread in descent anomalies over Central tropical Pacific during summer (R = −0.9; where R is the inter-model correlation, see “Methods”; Fig. 3b) and associated surface winds. This indicates that the inter-model spread in the winter Equatorial Pacific mainly arises from different tropospheric responses to the AMV forcing during summer. This inference is supported by the weaker inter-model correlation between winter NIÑO3.4 SST and winter Pacific descent responses (R = −0.64, not shown). To further understand the inter-model spread, we explore the origins of the tropical Pacific tropospheric descent anomalies in summer. Figure 3c shows that those subsiding anomalies are nearly fully mass-compensated by ascendant anomalies in other tropical regions. The 20°S–20°N tropical band (TROP) is further decomposed into a broad Indian ocean domain (TropInd), the Central Pacific ocean (TropPac), and a broad Atlantic domain (TropAtl; cf. indices definition in “Methods” and Fig. 2b). Through an analysis of variance (see “Methods”), we find that only 19% of the inter-model variance in TropPac descent anomalies is associated with the inter-model variance in TropAtl ascent anomalies, but 69% with the TropInd ascent ones. These two sources of spread are consistent with the two-phase mechanism detailed in the Introduction to explain the tropical Pacific response to Atlantic warming. In particular, it is consistent with the amplification of the Pacific response through the adjustment feedback of the Indo-Pacific Walker Circulation9. The key finding here is that there is no significant inter-model correlation between TropInd and TropAtl anomalies (R = 0.2, Fig. 3d). This indicates that models simulate different Indo-Pacific Walker Circulation adjustments (Phase-2 Indo-Pacific feedback) for similar Atlantic–Pacific atmospheric bridges (Phase-1 Atlantic forcing). Hence, this implies that the simulated feedback associated with the Indo-Pacific Walker Circulation adjustment is model-dependent and that the differences in this feedback are the source of most of the inter-model spread in the tropical Pacific response to the AMV forcing. We find two possible mechanisms to explain the different Indo-Pacific Walker Circulation adjustments among models in the AMV experiments. As detailed below, either different TropInd ascent or different TropPac SST responses can be the original driver of the different circulation responses. Further targeted experiments would be required to determine which mechanism is dominating here. However, both mechanisms point to the temperature response of the upper tropical troposphere as the key process to understand the inter-model differences: • The inter-model spread in TropInd ascent anomalies is tightly connected to the tropospheric lapse rate over the Warm Pool (Fig. 3e). Indeed, the larger the lapse rate (less warming in the upper troposphere compared to the surface), the less stable the troposphere is, and the more convectively active the tropical troposphere becomes. The inter-model spread in TropInd ascent anomalies is therefore linked to different responses in the upper tropospheric warming over the Warm Pool among models (Supplementary Fig. 4a–c). Because in the tropics the upper-tropospheric temperature is constrained by wave dynamical adjustment to be nearly horizontally uniform38,39 (Supplementary Fig. 4d), it implies that the TropInd ascent responses and the Indo-Pacific Walker Circulation responses are controlled by the different upper tropospheric warming among models. • The different upper tropospheric warming can lead to different SST warming among models through a “top-down” mechanism by a decrease of the surface latent heat flux40. There is indeed an inter-model correlation of R = 0.86 between the TROP upper tropospheric temperature and the TropPac SST responses (Fig. 3f). This “top-down” warming effect eventually modulates the amplitude of the TropPac descent and of the Indo-Pacific Walker Circulation adjustment. Therefore, for both the TropInd ascent and the TropPac SST mechanisms, the warmer the TROP upper troposphere is in response to an AMV warming, the weaker the Indo-Pacific feedback and the weaker the tropical wintertime Pacific cooling are. Then, in order to understand the inter-model spread in the wintertime Pacific response to AMV, one needs to understand the inter-model spread in TROP upper tropospheric temperature anomalies. As the thermal stratification of the tropical troposphere is primarily controlled by deep convection, upper tropospheric temperature anomalies in the Tropics can be generally traced to regional variations in the atmospheric boundary layer. To investigate the origins of those anomalies we study the moist static energy at the surface, using as estimate the equivalent potential temperature41 ($$\theta _{\mathrm{E}}$$; see “Methods”). In order to take into account the different contributions of highly active and less active convective regions in the injection of moist static energy from the surface to the upper troposphere, we weigh surface $$\theta _{\mathrm{E}}$$ with local precipitation (Pr in mm d−1) following Sobel et al.42’s approach (see “Methods”): $${\mathrm{P}}\theta _{\mathrm{E}} = a \times {\mathrm{Pr}} \times \theta _{\mathrm{E}}/ < a \times {\mathrm{Pr}} >$$ (1) Where $$< . >$$ symbols indicate the sum over TROP and $$a$$ is the grid cell surface area. The inter-model correlation between the changes of upper tropospheric temperature and our weighted $$\theta _{\mathrm{E}}$$ variable ($${\mathrm{P}}\theta _{\mathrm{E}}$$) summed over the tropical band is R = 0.96 (Fig. 4a), confirming the physical link between $${\mathrm{P}}\theta _{\mathrm{E}}$$ and the upper tropospheric conditions in the tropics. To further understand the origins of the inter-model spread, we decompose the $${\mathrm{P}}\theta _{\mathrm{E}}$$ anomalies into a term linked to precipitation anomalies only, i.e., $${\mathrm{P}}^\prime \theta _{{\mathrm{E}},{\mathrm{C}}}$$, a term linked to $$\theta _{\mathrm{E}}$$ anomalies only, i.e., $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}^\prime$$, and a covariance term, i.e., $${\mathrm{P}}_{\mathrm{C}}^\prime \theta _{\mathrm{E}}^\prime$$ (see “Methods”). We find that most of the inter-model spread in upper tropospheric temperature anomalies is coming from differences in the injection of surface moist static energy anomalies into the upper troposphere by the mean model vertical motions (the $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}^\prime$$ term; Fig. 4c). Furthermore, we find that the upper tropical troposphere warm anomalies are generated quasi-equally by anomalies occurring in the TropAtl and TropInd regions and, to a lesser extent, in TropPac (Fig. 4d). However, its inter-model spread is primarily driven by the TropAtl and TropPac sectors (black lines), with inter-model correlations between the upper troposphere temperature anomalies and $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}^\prime$$ summed over those regions equal to R = 0.96 and R = 0.87, respectively. Because the forcing is coming from the Atlantic in the present experiments, we assume that it is the spread in TropAtl $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}^\prime$$ that controls the spread in the tropical upper-tropospheric temperature and that the latter is amplified by the TropPac $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}^\prime$$ response. In summary, the analysis of $${\mathrm{P}}\theta _{\mathrm{E}}$$ indicates that the inter-model spread in the tropical upper tropospheric temperature anomalies can be explained by different injections of moist static energy from the TropAtl surface into the upper troposphere (Fig. 4b). This is eventually responsible for the modulation of the Indo-Pacific Walker Circulation feedback among models. Hence we identify two summertime variables centered over the TropAtl region that contribute to the inter-model spread in the tropical Pacific response: (1) the divergence of mass in the upper troposphere over TropAtl and (2) the injection of moist static energy anomalies from the TropAtl surface into the upper troposphere by the mean convective activity ($${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$). Building a bi-linear regression model with those two variables as predictors (see “Methods”), we capture as much as 73% of the inter-model variance in the wintertime NIÑO3.4 SST response (Fig. 5a, b); TropAtl ascent and $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$ accounting for 39% and 61% of the total regression model variance, respectively. ### Bias corrections and reduction of the uncertainty Next, we investigate the origins of the model response differences over TropAtl aiming at narrowing the uncertainty of our numerical estimate of the tropical Pacific response to the observed AMV forcing. We start by decomposing further the $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$ variable over the TropAtl region to evaluate whether its inter-model spread is coming from differences among models in climatological precipitation ($${\mathrm{P}}_{\mathrm{C}}^ \ast \left[ {\theta _{\mathrm{E}}^\prime } \right]$$), $$\theta _E$$ anomalies ($$\left[ {{\mathrm{P}}_{\mathrm{C}}} \right]\theta _{\mathrm{E}}^{\prime \ast }$$), or a combination of both ($${\mathrm{COV}}$$; see “Methods”). We find that all terms contribute to the inter-model spread, but that their respective importance is spatially dependent (Fig. 4e–h). Of particular interest, this analysis demonstrates that different climatological precipitation among models (Fig. 4f) is partly responsible for the inter-model spread. Because the model climatological precipitations are biased relative to observations, it implies that the simulated $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$ are also biased, which leads to erroneous estimates of the response to the observed AMV forcing. To minimize this error, we apply a bias correction to the $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$ of each model by computing them using the observed climatological precipitation instead of model one: $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$. This suppresses the spread of $${\mathrm{P}}_{\mathrm{C}}^ \ast \left[ {\theta _{\mathrm{E}}^\prime } \right]$$ but it introduces a new source of spread coming from observational uncertainties (see “Methods”). This bias correction decreases overall the inter-model variance of $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$ over TropAtl by 58%. Feeding our bi-linear NIÑO3.4 regression model with $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$ instead of $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$, we quantify that correcting for model mean precipitation biases helps to reduce the inter-model response variance over the tropical Pacific by 35% (Fig. 5b). Over the eastern Pacific (i.e., the western part of the wide TropAtl sector as shown in Fig. 4g), it is mainly the different $$\theta _{\mathrm{E}}$$ responses among models that drive the inter-model spread in $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$ and, a fortiori, in $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$ (Fig. 4g, see also Supplementary Fig. 11). $$\theta _{\mathrm{E}}$$ anomalies there are largely associated with surface temperature changes but their sign and amplitude are model-dependent (Supplementary Fig. 9), leading to compensating anomalies in the MMM (Fig. 2b). In the following, we demonstrate that the spread in $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$ over the eastern Pacific is explained by the different model climatological precipitations during February–March–April (Fig. 6g, h). During summer, all models simulate westerly anomalies north of 5°N associated with a northward shift of the Inter-tropical Convergence Zone (ITCZ) over the East Pacific in response to the AMV warming (Fig. 6c–f). Yet, by dividing the models into two sub-groups based on the state of their late winter climatological precipitation, we show that this shift is more pronounced for the models simulating a more northward position of the climatological ITCZ in late winter. Those models simulate an SST cooling around 7°N on the southern flank of the ITCZ in summer (Fig. 6a, b), where the other models simulate warming, which explains the inter-model spread in $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$. For the models with the largest ITCZ shift, we find that the westerly anomalies follow the seasonal migration of the precipitation anomalies; those are present north of the Equator since the winter when their cooling effect on the ocean is greatest (Fig. 6c, e). This suggests that preconditioning of the summertime cooling around 7°N during previous seasons occurs through feedback between ITCZ position, SST, wind, and surface flux anomalies43. Yet, all models tend to simulate a northward shift of the ITCZ in winter (Fig. 6c, d) but only some of them simulate such preconditioning. This inter-model disagreement is coming from different model climatological precipitation in late winter. For models simulating an ITCZ located north of the Equator, the northward shift of the ITCZ increases the mean south-westerlies and their associated turbulent heat fluxes around the Equator, which tends to cool locally the SST. On the other hand, for models simulating an ITCZ located South of the Equator, the northward shift of the ITCZ weakens the mean north-easterlies and their cooling effect on the equatorial SST. Given the high correlation between the climatological precipitation in February–March–April and the summertime $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$ response over East Pacific (R = −0.87; Fig. 6h), we use this information to further correct our estimate of the Tropical Pacific response to AMV. Associated with the observed February–March–April climatological precipitations, we estimate summertime East Pacific $$P_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$ values ranging from −0.09 °C and −0.13 °C (cf. green lines in Fig. 6h). Substituting these $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$ values for each model to the contribution of the East Pacific into the TropAtl $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$, we obtain TropAtl $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{{\mathrm{E}},{\mathrm{cor}}}^\prime$$ that we consider as our best estimate of the $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$ response to the observed AMV forcing (see “Methods”). Feeding our bi-linear regression model with $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{{\mathrm{E}},{\mathrm{cor}}}^\prime$$ instead of $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$, we quantify that correcting both summertime and late winter precipitation mean biases reduces by 65% the inter-model variance in our analytical estimate of the NIÑO3.4 response (Fig. 5b). We also investigated the potential origin of the inter-model spread over Atlantic–Africa (i.e., the eastern part of the TropAtl region, cf. Fig. 4e–h). We found that the inter-model spread in $$\left[ {{\mathrm{P}}_{\mathrm{C}}} \right]\theta _{\mathrm{E}}^{\prime \ast }$$ anomalies is associated with different signs in the SST and specific surface humidity responses around the eastern equatorial Atlantic (Supplementary Figs. 3 and 4). However, we did not identify the physical processes controlling the different model behaviors (cf. Supplementary Discussion). ## Discussion Using 21 coordinated simulations from 13 different CGCMs, we show that: • In response to an AMV warming, all models simulate tropical Pacific changes reminiscent of La Niña conditions. This result confirms the influence of the Atlantic on climate variability at the global scale and it supports the idea that the AMV has contributed to the 1998–2012 global warming slowdown through its impacts on the tropical Pacific. However, the strength of the connection varies by a factor of 10 between the models. • The tropical Pacific response to the Atlantic forcing is driven by changes in (1) the Atlantic–Pacific Walker Circulation and (2) the amount of moist static energy injected from the Atlantic surface into the upper troposphere. • The latter is responsible for most of the uncertainty in our current numerical model estimates of the Pacific response to the observed AMV, mainly because of different mean precipitation climatology. Partially correcting for mean model precipitation biases, we reduce this uncertainty and we specifically quantified that the NIÑO3.4 response to an observed 0.26 °C AMV warming ranges from −0.05 °C to −0.16 °C with a median value of −0.11 °C and an inter-model standard deviation of 0.03 °C. We acknowledge that this estimate is still subject to model limitations. In particular, we reduce uncertainty by correcting a posteriori for model precipitation biases. Any possible interactions between those biases and surface equivalent potential temperature responses to AMV would still affect our estimate. Therefore, our analysis highlights the importance of reducing mean climate model biases in order to properly simulate and predict the global AMV impacts. Although this study focuses on decadal timescales signals, the discussed mechanisms take place at monthly timescales. Our study shows then the potential for improving climate predictions from seasonal to decadal timescales through a better representation of the impacts of the Atlantic on tropical Pacific28,29,44. The discussed mechanisms very likely also act to shape the Pacific mean state and their differences among models45,46,47, which are partly responsible for the inter-model spread in climate projections48. Based on our findings, we suggest that the analysis of the injection into the upper troposphere of moist static energy from the Atlantic surface can be used as an interpretative framework to understand the inter-model uncertainties around future climate simulations. Finally, we note that several observational and model-based studies49,50,51 suggest the existence of a two-way interaction between Atlantic and Pacific at decadal timescale: an AMV warming driving a Pacific cooling, which eventually drives an Atlantic cooling. Due to the experimental protocol used in the present article, we could only focus on the representation by models of the Atlantic impacts on the Pacific. To persist in exploring the sources of climate predictability at multi-annual timescale and their current limits due to model uncertainty, a similar multi-model study to this one should be completed but investigating the Pacific impacts on the Atlantic. ## Methods ### Experiments The 21 experiments from 13 different CGCMs used in this study are listed in Supplementary Tables 1 and 2; it represents a total of 12,320 simulated years. Following the DCPP-C protocol23, two sets of ensemble simulations have been performed for each experiment, in which time-invariant SST anomalies corresponding to the warm (AMV+) and cold (AMV−) phases of the observed AMV were imposed over the North Atlantic using SST nudging. To capture the potential response and adjustment of other oceanic basins to the AMV anomalies, the simulations were integrated for 10 years with fixed external forcing conditions. Large ensemble simulations were performed in order to robustly estimate the climate impacts of the AMV (from 10 to 50 members depending on the model, cf. Supplementary Table 2). An extensive description of the experimental protocol is provided in the Technical note for AMV DCPP-C simulations: https://www.wcrp-climate.org/wgsip/documents/Tech-Note-1.pdf. Over the North Atlantic (Equator-65°N/80°W–0°), the spatial correlation of the SST anomalies in each simulation and the observed AMV target varies between 0.66 and 0.86, with a multi-model average value of 0.79, indicating that all simulations are constrained by similar SST conditions in the North Atlantic. We note that the idealized AMV simulations underestimate by ~20% the amplitude of the observed AMV target. This is because we do not impose a very strong nudging in the experimental protocol to allow ocean-atmosphere coupling and variability at high frequency (as recommended by the CMIP6/DCPP-C protocol23), which tends to dissipate the heat anomalies imposed at the surface. Further evaluation of the experimental protocol is provided in Supplementary. Some simulations deviate from the AMV DCPP-C protocol. CESM1 simulations used an observed AMV pattern computed from the ERSSTv3b dataset52 instead of ERSSTv453. CNRM-CM6-1-HR, EC-Earth3P-HR, EC-Earth3P-LR, ECMWF-IFS-HR, and ECMWF-IFS-LR used a constant 1950 or 1990 (instead of 1850) external forcing background (cf. Supplementary Table 2). The impact of the external forcing background on the results is tested with the CNRM-CM5 models for which AMV simulations have been performed with both 1850 and 1990 backgrounds. We did not find evidence for the impact of the protocol differences on the results discussed in this article. In addition, the MetUM-GOML-HR and MetUM-GOML-LR simulations used a 1000-m mixed-layer ocean model and 1990 external forcing background. Those models offer insights on the role played by the ocean dynamics in the documented climate responses when compared to the models with full ocean dynamics. Finally, the imposed AMV forcing strength is not the same for all simulations. As detailed in the column “AMV strength” of Supplementary Table 2, the imposed AMV anomalies vary between 1, 2, and 3 times the observed AMV standard deviation. Assuming linearity in the AMV responses, we weight each simulation by dividing their output by the AMV forcing strength in order to compare the results from all the AMV experiments. This is done for all the figures in the article. This enables us to create a larger multi-model ensemble and to evaluate more precisely the origins of the inter-model spread. Scaled outputs from experiments performed with the same model but with different AMV strengths are often indistinguishable, which suggests that the linear assumption is a reasonable approximation for the analyses of this study. Yet, we highlight the different AMV strengths by different colors in the figures (1×AMV: blue; 2×AMV: orange; 3×AMV: magenta). ### MMM and inter-model correlations (R) The MMM is computed by averaging the ensemble mean of each simulation, regardless of the number of ensemble members (i.e, there is no weighting). The outputs of each simulation are scaled by their AMV strength forcing prior to computing the MMM (as described above). For models for which several sets of experiments have been performed (with different magnitudes of AMV anomalies and/or different external forcing backgrounds), we average all the experiments of each model together prior to computing the MMM in order to not bias the results toward an over-represented model (e.g., CNRM-CM5 or EC-Earth3P). Because of the absence of ocean dynamics in the two MetUM-GOML models, those models are not taken into account in the computation of the MMM. Similarly to the computation of the MMM, the inter-model correlation R is computed after averaging all the ensemble means from the same model (if more than one experiment was performed) in order to give the same weight to all models. We also computed the inter-model correlation based on all the ensemble means from all the simulations (i.e., no averaging of experiments from the same model prior to the computation of the correlation) but no significant differences between the two correlations were found for the relationship investigated in this article. Because of the absence of ocean dynamics in the two MetUM-GOML models, those models are not taken into account in the computation of inter-model correlations. ### Regions definition To assess the tropical Pacific response, we use the NIÑO3.4 index defined as the SST averaged over 5°S–5°N and 170°W–120°W (Fig. 2a). Based on the summer MMM anomalies of precipitation and vertical velocity at 500hPa (Supplementary Fig. 3e, f), we decomposed the 20°S–20°N tropical band (TROP) into three main regions: a broad Indian region spanning from 30°E to 135°E, a central Pacific region spanning from 135°E to 120°W, and a broad Atlantic region spanning from 120°W to 30°E (Fig. 2b). We label those regions TropInd, TropPac, and TropAtl, respectively. In addition, an East Pacific and an Atlantic–Africa regions (embedded into TropAtl) are used in Figs. 4 and 6 that cover 120°W–80°W/20°S–20°N and 80°W–30°E/20°S–20°N, respectively. ### Analysis of variance Taking advantage of the quasi-mass compensation of the vertical motion in the TROP region (Fig. 3c), we estimate the origins of the inter-model spread in TropPac descent through an analysis of variance: $$S_{{\mathrm{TropPac}}}^2 \sim S_{{\mathrm{TropAtl}} + {\mathrm{TropInd}}}^2 = S_{{\mathrm{TropAtl}}}^2 + S_{{\mathrm{TropInd}}}^2 + {\mathrm{COV}}$$, where $$S_{{\mathrm{TropPac}}}^2$$, $$S_{{\mathrm{TropAtl}}}^2$$ and $$S_{T{\mathrm{ropInd}}}^2$$ are the inter-model variance in TropPac, TropAtl, and TropInd descent anomalies, respectively; $${\mathrm{COV}}$$ is the covariance term between TropAtl and TropInd descent anomalies and $$S_{{\mathrm{TropAtl}} + {\mathrm{TropInd}}}^2$$is the inter-model variance of descent anomalies averaged over the whole Trop region excluding the TropPac region. We find that $$S_{{\mathrm{TropAtl}}}^2$$, $$S_{{\mathrm{TropInd}}}^2$$ and COV explains 19%, 69%, and 12% of $$S_{{\mathrm{TropAtl}} + {\mathrm{TropInd}}}^2$$, respectively. ### Equivalent potential temperature $$\theta _{\mathbf{E}}$$ Theoretically, the equivalent potential temperature can be defined as $$\theta _{\mathrm{E}}\sim \theta exp\left( {\frac{{L_{\mathrm{C}}q_{\mathrm{S}}}}{{C_PT}}} \right)$$, where $$\theta$$ is the dry potential temperature, $$L_{\mathrm{C}}$$ is the latent heat of condensation, $$q_{\mathrm{S}}$$ is the saturation of mixing ratio, $$C_{\mathrm{P}}$$ is the specific heat of dry air, and $$T$$ is the temperature. This formula explicitly shows that $$\theta _{\mathrm{E}}$$ is similar to the potential temperature for dry air mass (which remains constant during adiabatic processes) but it corrects for the energy associated with the air mass moisture, assuming that all the energy released by condensation/evaporation remains in the air mass (pseudo-adiabatic process). Here we used the NCL function “pot_temp_equiv_tlcl” (https://www.ncl.ucar.edu/Document/Functions/Contributed/pot_temp_equiv_tlcl.shtml) to compute $$\theta _E$$. This function is based on Eq. (39) from Bolton54, which gives more accurate results than the theoretical formula given above but that requires the computation of the temperature at the lifted condensation level. Such temperature is estimated with the NCL function “tlcl_rh_bolton” (https://www.ncl.ucar.edu/Document/Functions/Contributed/tlcl_rh_bolton.shtml), which is based on Eq. (22) from Bolton54. ### Weighted equivalent potential temperature $${\mathbf{P}}\theta _{\mathbf{E}}$$ as a proxy for the upper-tropospheric temperature Over the oceans, the mean tropospheric temperature profile is often considered to be in a moist-adiabatic convective equilibrium with the mean SST55 (Supplementary Fig. 4f), as the SST controls directly the atmospheric boundary layer energy content. Yet, convective adjustment can act directly only in regions of frequent precipitation, which are mostly over warm SST regions. In regions of no convection, the surface has no direct means of influencing the free troposphere, and the SST anomalies cannot shape the tropospheric temperature profile. Hence, there is no evident physical reason for considering mean tropical SST variations as a proxy for upper tropospheric temperature anomalies. To take into account the different contributions of local SST to the upper-tropospheric temperature, Sobel et al. 42 introduced a more appropriate proxy by weighting the SST with the local precipitation before computing the tropical average. We follow this method here, but we generalize it in order to account for the effect of deep convection overland on the upper tropospheric temperature anomalies56 and we compute the precipitation weighted equivalent potential temperature $${\mathrm{P}}\theta _{\mathrm{E}}$$, cf. Eq. (1). ### Notation We note $$<\, f \,>$$ as the sum of the values of a given field $$f$$ for all tropical grid points within 20°S–20°N. We define $$f = f_{\mathrm{C}} + f\prime$$, where $$f\prime$$ is the departure of $$f$$ from $$f_{\mathrm{C}}$$, which is the time-averaged ensemble mean of the AMV- experiments (the AMV- experiment being considered as the reference state). We also define $$= \left[ f \right] + f^ \ast$$, where $$f^ \ast$$ is the departure of $$f$$ from its multi-model mean $$\left[ f \right]$$. ### Decompositions of $${\mathbf{P}}\theta _{\mathbf{E}}$$ In the article, $${\mathrm{P}}\theta _{\mathrm{E}} = \frac{{a \times {\mathrm{Pr}} \times \theta _{\mathrm{E}}}}{{ < a \times {\mathrm{Pr}} > }}$$ is first decomposed into a term linked to precipitation anomalies only $${\mathrm{P}}\prime \theta _{{\mathrm{E}},{\mathrm{C}}} = \frac{{a \times {\mathrm{Pr}}\prime \times \theta _{{\mathrm{E}},{\mathrm{C}}}}}{{ < a \times {\mathrm{Pr}}\prime > }} - \frac{{a \times {\mathrm{Pr}} \times \theta _{\mathrm{E}}}}{{ < a \times {\mathrm{Pr}} > }} \times \frac{{a \times {\mathrm{Pr}} \times \theta _{\mathrm{E}} \times {\mathrm{Pr}}\prime }}{{ < a \times {\mathrm{Pr}} > }}$$, a term linked to $$\theta _{\mathrm{E}}$$ anomalies only $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime = \frac{{a \times {\mathrm{Pr}}_{\mathrm{C}} \times \theta _{\mathrm{E}}\prime }}{{ < a \times {\mathrm{Pr}}_{\mathrm{C}} > }}$$, and a covariance term $${\mathrm{P}}\prime \theta _{\mathrm{E}}\prime = \frac{{a \times {\mathrm{Pr}}\prime \times \theta _{\mathrm{E}}\prime }}{{ < a \times {\mathrm{Pr}}\prime > }}$$. In a second time $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$ is decomposed into a term linked to the climatological precipitation differences among models $${\mathrm{P}}_{\mathrm{C}}^ \ast \left[ {\theta _{\mathrm{E}}^\prime } \right] = \frac{{a \times {\mathrm{Pr}}_{\mathrm{C}}^ \ast \times \left[ {\theta _{\mathrm{E}}\prime } \right]}}{{ < a \times {\mathrm{Pr}}_{\mathrm{C}}^ \ast > }}$$, a term link to the different $$\theta _{\mathrm{E}}$$ response to AMV among models $$\left[ {P_{\mathrm{C}}} \right]\theta _{\mathrm{E}}^{\prime \ast } = \frac{{a \times \left[ {{\mathrm{Pr}}_{\mathrm{C}}} \right] \times \theta _{\mathrm{E}}^{\prime \ast }}}{{ < a \times \left[ {{\mathrm{Pr}}_{\mathrm{C}}} \right] > }}$$, and a covariance term $${\mathrm{COV}} = \frac{{a \times {\mathrm{Pr}}_{\mathrm{C}}^ \ast \times \theta _{\mathrm{E}}^{\prime \ast }}}{{ < a \times {\mathrm{Pr}}_{\mathrm{C}}^ \ast > }}$$. ### Bi-linear regression model The coefficient of the bi-linear regression model is computed using as predicand the wintertime NIÑO3.4 SST index and as the two predictors the summertime vertical ascent summed over TropAtl and the summertime $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$ summed over TropAtl. The two MetUM-GOML simulations are excluded from the computation of the regression model coefficients and, similarly, as for the inter-model correlation, all models share the same weight. ### Bias corrections and observational uncertainties Two bias corrections are applied to $${\mathrm{P}}_{\mathrm{C}}\theta _{\mathrm{E}}\prime$$. First, we compute this variable by using observed climatological precipitation ($${\mathrm{Pr}}_{{\mathrm{Obs}}}$$) instead of model climatological precipitation: $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime = \frac{{a \times {\mathrm{Pr}}_{{\mathrm{Obs}}} \times \theta _{\mathrm{E}}\prime }}{{ < a \times {\mathrm{Pr}}_{{\mathrm{Obs}}} > }}$$. Then, we decomposed the TropAtl $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$ into its regional components coming from the Atlantic–Africa region, the East Pacific region, and their covariance term. The East Pacific component is then substituted by a value estimated from the observed climatological precipitation and the inter-model relationship between JJAS East Pacific $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$ and February–March–April climatological precipitation centroid over East Pacific (Fig. 6h). Following this substitution, we sum again the different components of TropAtl $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$to obtain $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{{\mathrm{E}},{\mathrm{cor}}}^\prime$$. In order to account for observational uncertainties57,58, we compute for each model five $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{\mathrm{E}}\prime$$and $${\mathrm{P}}_{{\mathrm{Obs}}}\theta _{{\mathrm{E}},{\mathrm{cor}}}^\prime$$ values using different observation estimates. ### Observational and reanalysis datasets The SST from the ERSSTv453 and ERSSTv3 datasets52 were used to extract the observed AMV pattern imposed in the simulations. The HadCRUT459 data set was used to compute the observed AMV composites shown in Fig. 1c. CMAP60,61, GPCPv2.362, TRMMv7 at 0.5° of spatial resolution63,64, MSWEPv2.665, and ERA-Interim66 data sets were used for mean bias corrections of model precipitation (cf. Figs. 5 and 6). Depending on data availability, we used different periods to compute the observed estimate mean state: 1979-2017 for CMAP, GPCPv2.3, and MSWEPv2.6; 1998-2011 for TRMMv7; and 1979–2018 for ERA-Interim.
2023-02-01 07:07:27
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 2, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5710663795471191, "perplexity": 2727.098686959111}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499911.86/warc/CC-MAIN-20230201045500-20230201075500-00103.warc.gz"}
https://puzzling.stackexchange.com/questions?sort=newest&page=2
All Questions 16,967 questions 2answers 84 views Time detective duty Consider that you have a beautiful wife (husband). But your after-marriage life is not so beautiful. Your wife (husband) has performed a first adultery act 5 years after a marriage with you and later ... 1answer 48 views What's the solution of this Raven's matrix? (2) [duplicate] here's another matrix which I don't understand: What's the solution and why? Thanks! 1answer 1k views Can you cover a cube with copies of this shape? The following shape has an interesting property: It is possible to map multiple copies of this shape onto the surface of a cube in a way that perfectly covers the entire cube with no gaps and no ... 3answers 1k views Catching a robber on one line At x = 0, a thief robbed a bank. The thief ran one of two known directions at a constant speed, towards x < 0 or towards x > 0. The cop arrives at the crime scene some unknown time after the ... 2answers 161 views Preparing for the End - Riddle This is the end. I knew it was coming. All of my ends are tied up, I'm ready. Only a few more minutes to spend. I blame you for this. You killed the messenger. I strained and cried ... 0answers 53 views Distinct SemiPrimes with Close Factors I'm in the midst of developing a game for several friends as a friendly competition. It begins with generating the largest distinct semiprime feasible with factors that closely related in length ... 4answers 174 views Problem with heights - Riddle I hate heights. As soon as I can see how far away the ground is, I'm filled with this uncontrollable fear and urge to get down -any survivable way to get down is better than this. My captors seem ... 0answers 65 views 1answer 114 views Rhyme with Time Commonly used units of Time are: $Second$ $Minute$ $Hour$ $Day$ $Week$ $Month$ $Year$ $Decade$ $Century$ $Millenium$ Can you find at least one rhyming word in English Language for each of ... 4answers 3k views Fastest way from 10 to 1 with everyone in between I got inspired by this question from trolley813 to make this puzzle. How fast one can get from TEN to ONE, passing each number in between? The rules are simple: You are only allowed to change ... 1answer 133 views Fill.in the two missing terms in this Simple Sequence What are the two missing numbers in the sequence shown below: \def\S{\phantom{XX}} \large\begin{array}{|c|c|c|c|c|c|}\hline \S&\S&\S&\S&\S&\S\\[-10px] 4 & 5 & 6 & 7 &... 1answer 128 views What's the solution of this Raven's matrix? I'm practicing with Raven's Matrices but I don't understand this one. What's the solution and why? Thanks! 1answer 6k views How can this shape perfectly cover a cube? The following shape: can be folded onto the surface of a cube in a way that perfectly covers the entire cube with no gaps and no overlaps. How can this be done? 0answers 68 views Can this Rubik's cube be solved? [duplicate] All sides of the Rubik's cube are solved, except the corners. And these corners need to be swapped. 1answer 92 views How to swap adjacent corners? I want to swap adjacent corners of U layer, the only thing I care is: position of other 6 corners. I don't care: orientation of all corners position of all edges orientation of all edges The goal ... 5answers 273 views A Humourous & Simple Selfmate In One Here's simple and humourous chess puzzle that I created after a few non-consecutive hours. I had to prove the concept feasible while making it as economic as possible. White To Move, And Selfmate In ... 1answer 60 views Musical Chairs Cipher: Intermediate This question is a continuation of the first question: Musical Chairs Cipher. The basic rules: Six children -- Alice, Ben, Carl, Denise, Eddie, and Flo -- are playing a game of musical chairs in ... 2answers 235 views A Device Riley Riddle My prefix may be what a device can be on My infix is most likely in your device My suffix is what devices can be My whole is a reverse motion Hint: 0answers 19 views Upside down A (for all) [migrated] I am working on a voltmeter and the tolerance for it is 0.25% upside down A (for all) 1 digit. I am not sure what exactly what this means. I am guessing if my value was 19.95 'for all' could mean I ... 4answers 234 views Savage Road Signs (Part 2) Please read part 1 or this might be confusing Since part 1, you have replaced the stolen stickers and your daughter has forgiven you. The highway ended up being a full 700km long, so you are happy ... 0answers 105 views Don't even try, no need [closed] Here is the a resume of this post: Downvotes on this post, Once you get to the end of this riddle you will get it, Am I playing with words? Nah, I'm making everyone look stupid, Some of you will be ... 2answers 85 views Lunatic Loves this Sunny Sequence…Find the Missing info Complete this Sunny Sequence, Based on the information given below: 1answer 132 views Change EDIT to UNDO On your computer right now, you can probably click on the "Edit" menu; and one of the first options will be "Undo". These are closely associated words. But how close are they in a word ladder? ... 3answers 183 views 16 people, make 3 consecutive round of 4 players I am struggling with this issue for a tournament I am planning. Take 16 people. Consider a 4-player game. To be clear, in each round all 16 players must play the game once. I want to make at least 3 ... 1answer 653 views I am caught when I was about to steal some candies I am a eight years old boy, I love candies, but my mom does not want me to eat any more candies because I start to have black teeth :( but I still love candies so... Today when I was about to steal ... 6answers 4k views Can you open the door or die? v2 This is the follow up question to my previous question You wake up and find yourself in a very strange room with a single door. You find a note on the ground surrounded by lots of cards: ... 2answers 108 views Pan Digital Lucky Seven wants you to figure out all the digits $Given$: A, B, C, D, P, Q, R, S, T are distinct digits varying from 1 to 9. AB, CD, PQ, RST are all concatenated Numbers. From the given expression , figure out all the digits to make lucky seven. ... 5answers 4k views Fastest way from 8 to 7 Inspired by this great question by Uvc (actually, it's the same logic but with different words). How fast one can get from EIGHT to SEVEN? The rules are the same: You are only allowed to change one ... 3answers 214 views What are the missing members in this sequence based on the digits 0 to 9? Find the missing members? 0, 0, 2, 0, 2, 3, 0, ?, ?, ? Hint 1 4answers 2k views Savage Road Signs There is a highway that starts in the city of Savage. You must must place distance marker signs on this highway for the outgoing traffic. According to highway code, there must be a distance marker ... 4answers 2k views Can you open the door or die? You wake up and find yourself in a very strange room with a single door. You find a note on the ground surrounded by lots of cards: There are 66 green and 66 blue door cards on the ground. If you ... 1answer 62 views CiSRA Puzzle Hunt Backups? I found this article describing what makes a good puzzle, and it references several puzzles which, as best as I can tell, were designed for something called the CiSRA Puzzle Hunt. However, the website ... 1answer 75 views Primes and Puzzles are my Passion. Combining them here to get Productive Prime Puzzles..Find the Missing Info $EXAMPLE$: $997$ = $O.$ $T.$.- $T.$ One Thousand - Three $Expand$ on the Full Info of the following Prime clues. A). 2 = F & O. E. P B). 3 = S.P C). 5 = F. On the H. D). 7 = H. of ... 1answer 166 views Stars of the Celestial Bagua Check this out: Okay, it might be more impressive if I show you what is moving along the arrows: At every vertex in a long word. Flowing in to every vertex are two short words ("inputs"). ... 15 30 50 per page
2019-06-18 07:46:17
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5947438478469849, "perplexity": 1733.0307622576706}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560627998690.87/warc/CC-MAIN-20190618063322-20190618085322-00043.warc.gz"}
https://docs.siren.io/10.3.3/platform/en/siren-ml/data-sampling-guidelines.html
# Siren Platform User Guide ### Data sampling guidelines Data used for model training should be representative of the data as a whole, containing the scope of the values and all the patterns that are typical in the dataset. This maximizes the accuracy for detection and prediction models. You should carefully consider the sampling resolution and the range of the data when training a model to balance this against the efficiency of training. #### Bucket size The bucket size is the resolution that your data is sampled at both during training and for detections/predictions. The bucket size should be at least as large as the typical rate that your data is logged at. This means that if data is being logged at hourly intervals, the minimum bucket size value should be one hour. If you choose a lower value, your machine learning model will be trained on missing data; therefore, it will waste a lot of computational power learning trends that are not useful or that do not really exist. In many cases, choosing a larger bucket size can maintain the characteristics of the data while reducing the amount for training, speeding up model creation. Use the preview graph in the model creation screen as an indicator of a good bucket size and time range. As a practical example, let's say you want to train a model on a year's data which was logged at one-second intervals. A bucket size of 1 second yields 31.5 million points, whereas a bucket size of 5 minutes reduces this to 100k points. The preview graph in the model creation screen can again give an indication of a good bucket size and time range. #### Seasonality When training a model, you should ensure that the training range captures the seasonal changes within your data (seasonal variations occur at specific regular intervals of less than a year). For example, if your data has different patterns on weekdays and weekends, the training range should be over multiple weeks. If your data changes between summer and winter, use multiple years of data to accurately capture this trend. #### False positive anomalies An anomaly detection model may mark as anomalies some data points that you may consider normal. To rectify this, retrain the model, making sure to include the data that it incorrectly tagged as anomalous. #### Use cases Coastal temperature An oceanographer with a dataset measuring coastal sea temperature at a depth of 15m every day wishes to predict changes in the temperature signal. A model trained on the last two decades (7,300 points) would more accurately cover the ongoing sea temperature changes to allow more accurate prediction. Web server response time Anomaly detection models work very well for server logs. An example of this is response logs for a web server: { "time": "17/May/2018:08:05:32 +0000", } For this scenario, you would create an anomaly detection model for the saved search that has the logs, looking at the max of response_time. Depending on how many requests the web server gets, bucket size might be somewhere between 5 seconds (250k documents) and 1 minute (20k documents). Several weeks of data should be selected, which is enough to see variations over the weekends.
2020-04-01 17:03:40
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.366384893655777, "perplexity": 1053.2720978852994}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370505826.39/warc/CC-MAIN-20200401161832-20200401191832-00494.warc.gz"}
http://www.whoinventedfirst.com/tag/electromagnetism/
## Who invented the Maxwell’s Equations Importance of Maxwell’s Equations. Brilliant Electromagnetic Synthesis of Electrical & Magnetic Effects. Maxwell’s equations synthesized physicists’ understanding of electrodynamics and predicted the existence of electromagnetic waves.... ## Who discovered the Electromagnetic Effects Hans Christian Oersted’s Electromagnetic Effects. Oersted Discovers Electric Currents Induce Magnetic Fields. During a class, Oersted noticed that a compass near a wire carrying electricity was affected... ## Who discovered the Electromagnetism Life and works of James Clerk Maxwell, considered the 19th century’s greatest scientist, famous for electromagnetism and Maxwell’s Equations. James Clerk Maxwell Biography. Scottish Physicist, Arguably Greatest... ## Andre-Marie Ampere French Physicist and Mathematician Brief biography of French scientist André-Marie Ampère who laid the foundation of electromagnetism, he called electrodynamics. André-Marie Ampère is best known for his significant contributions to...
2022-11-30 06:55:43
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8649023175239563, "perplexity": 5990.8185430014155}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-49/segments/1669446710733.87/warc/CC-MAIN-20221130060525-20221130090525-00642.warc.gz"}
https://tex.stackexchange.com/questions/590934/why-does-an-empty-tabularx-throw-an-error
# Why does an empty tabularx throw an error? Why does an empty tabularx throw an error ? (by empty, I mean only space, no newline) MWE: \documentclass{article} \usepackage{tabularx} \begin{document} \begin{tabular}{|c | c|} \end{tabular} % <-- Works \begin{tabular}{|c | c|}\end{tabular} % <-- Works \begin{tabularx}{\textwidth}{|c | c|} \end{tabularx} % <-- Works \begin{tabularx}{\textwidth}{|c | c|}\end{tabularx} % <-- Does not work \end{document} (This is particulary frustrating with \ExplSyntaxOn) • "why" is probably that it didn't occur to me to test for this and trap it. – David Carlisle Apr 1 at 9:37 • Well... I guess it's probably a good reason ^^' – hl037_ Apr 1 at 9:54 • This code has been available for almost 30 years and you are the first person to report this issue:-) – David Carlisle Apr 1 at 9:56 "why" is probably that it didn't occur to me to test for this and trap it. Looking at tracing all the reason is that in \begin{tabularx}{\textwidth}{|c | c|}\end{tabularx} The entire body (that tabularx needs to grab as a macro argument) is {|c | c|} and tex silently drops brace groups that surround an entire delimited argument, so this is passed to tabularx as \begin{tabularx}{\textwidth}|c | c|\end{tabularx} and things go wrong. If you are generating this and there is a chance it might be empty you could use the optional argument \begin{tabularx}{\textwidth}[c]{|c | c|}\end{tabularx} % <-- Does not work works for example. I can probably catch this in a future release. possible fix \documentclass{article} \usepackage{tabularx} \makeatletter \def\foo#1{% \def\tabularx##1{% \edef\TX@{\@currenvir}% {\ifnum0=}\fi \setlength\TX@target{##1}% \TX@typeout{Target width: ##1 = \the\TX@target}% \toks@{}\TX@get@body#1}} \foo{ } \makeatother \begin{document} \begin{tabularx}{\textwidth}{|c | c|}\end{tabularx} % <-- Does not work \end{document} ` • In my case, it's not that a big deal once i know it can't be empty. I actually encountered it when debugging a macro to generate table, and I stuck on this error not knowing that an empty tabularx would error – hl037_ Apr 1 at 9:57 • @hl037_ see update – David Carlisle Apr 1 at 10:53
2021-07-30 17:21:13
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7971144318580627, "perplexity": 4515.707761562569}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046153971.20/warc/CC-MAIN-20210730154005-20210730184005-00313.warc.gz"}
http://www.mathnet.ru/php/archive.phtml?wshow=paper&jrnid=nd&paperid=130&option_lang=eng
Nelineinaya Dinamika [Russian Journal of Nonlinear Dynamics] RUS  ENG JOURNALS   PEOPLE   ORGANISATIONS   CONFERENCES   SEMINARS   VIDEO LIBRARY   PACKAGE AMSBIB General information Latest issue Archive Impact factor Search papers Search references RSS Latest issue Current issues Archive issues What is RSS Rus. J. Nonlin. Dyn.: Year: Volume: Issue: Page: Find Nelin. Dinam., 2007, Volume 3, Number 2, Pages 123–140 (Mi nd130) Gibbs ensembles, equidistribution of the energy of sympathetic oscillators and statistical models of thermostat V. V. Kozlov Steklov Mathematical Institute, Russian Academy of Sciences Abstract: The paper develops an approach to the proof of the “zeroth” law of thermodynamics. The approach is based on the analysis of weak limits of solutions to the Liouville equation as time grows infinitely. A class of linear oscillating systems is indicated for which the average energy becomes eventually uniformly distributed among the degrees of freedom for any initial probability density functions. An example of such systems are sympathetic pendulums. Conditions are found for nonlinear Hamiltonian systems with finite number of degrees of freedom to converge in a weak sense to the state where the average energies of the interacting subsystems are the same. Some issues related to statistical models of the thermostat are discussed. Keywords: Hamiltonian system, sympathetic oscillators, weak convergence, thermostat Full text: PDF file (263 kB) UDC: 530+517 MSC: 37A60, 60K35, 70H05, 82B30, 40A99 Citation: V. V. Kozlov, “Gibbs ensembles, equidistribution of the energy of sympathetic oscillators and statistical models of thermostat”, Nelin. Dinam., 3:2 (2007), 123–140 Citation in format AMSBIB \Bibitem{Koz07} \by V.~V.~Kozlov \paper Gibbs ensembles, equidistribution of the energy of sympathetic oscillators and statistical models of thermostat \jour Nelin. Dinam. \yr 2007 \vol 3 \issue 2 \pages 123--140 \mathnet{http://mi.mathnet.ru/nd130} • http://mi.mathnet.ru/eng/nd130 • http://mi.mathnet.ru/eng/nd/v3/i2/p123 SHARE: Citing articles on Google Scholar: Russian citations, English citations Related articles on Google Scholar: Russian articles, English articles This publication is cited in the following articles: 1. V. V. Kozlov, “The generalized Vlasov kinetic equation”, Russian Math. Surveys, 63:4 (2008), 691–726 2. A. V. Kargovsky, L. S. Bulushova, O. A. Chichigina, “Theorem on energy distribution over degrees of freedom for a quasistable symmetric anharmonic oscillator”, Theoret. and Math. Phys., 167:2 (2011), 636–644 3. A. P. Kuznetsov, S. P. Kuznetsov, L. V. Tyuryukina, I. R. Sataev, “Stsenarii Landau–Khopfa v ansamble vzaimodeistvuyuschikh ostsillyatorov”, Nelineinaya dinam., 8:5 (2012), 863–873 • Number of views: This page: 224 Full text: 57 First page: 1
2022-01-22 12:42:43
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.20905174314975739, "perplexity": 6948.6581380522075}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320303845.33/warc/CC-MAIN-20220122103819-20220122133819-00044.warc.gz"}
https://aviation.stackexchange.com/questions/93278/what-is-the-difference-between-unserviceable-unusable-and-not-maintained
What is the difference between unserviceable, unusable, and not maintained? Here are some NOTAMS for a recent flight I was planning: LXY 03/002 LXY NAV NDB U/S 2203081935-PERM CRS 02/008 CRS NAV NDB UNUSABLE 2202141320-2206142111EST ANB 05/580 SJI NAV VOR NOT MNT 2205201953-2205232000EST We all see these kinds of NOTAMS all the time. In each case, I would consider each NAVAID to be unfit for navigational planning or flight use purposes, meaning that I now consider all of them out of service and not to be used. But is there any actual semantic difference between the qualifier in these three cases? Or do they all just mean, "it's not working, don't use it!"? If so, why are three separate qualifiers used when UNUSABLE would be sufficient?
2023-03-20 15:26:50
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.404225617647171, "perplexity": 4315.312862132803}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296943484.34/warc/CC-MAIN-20230320144934-20230320174934-00326.warc.gz"}
https://physics.stackexchange.com/questions/452560/are-there-changing-magnetic-and-electric-fields-that-are-not-em-radiation
# are there changing magnetic and electric fields that are not EM radiation? Let us consider these two Maxwell equations: $$\frac{\partial \vec{B}}{\partial t}=-\vec{\nabla}\times \vec{E}$$ and $$\frac{\partial \vec{E}}{\partial t}=\frac{1}{\epsilon_0}\left(-\vec{J}+\frac{1}{\mu_0}\vec{\nabla}\times \vec{B}\right).$$ When we consider faraday's law of induction, we usually assume that the changes are slow, and thus we can neglect radiation by assuming that the left hand side of the second equation is zero. That is, a changing current creates a changing magnetic field, which in turn creates a changing electric field, per the first equation. I cannot understand this. First, if we can neglect the change in E from the second equation, should not we also neglect the change in B in the first equation? Second, this imply that we can have changing electric and magnetic fields that are not electromagnetic waves. But are not all changing magnetic or electric fields EM waves? or is this approximation equivalent to a charge moving at constant speed, in which the change in E and B are not due to radiation but just to the translation motion of the static field lines An electric or magnetic field always obeys a wave equation this can be proven by eliminating one or the other from the two equations that you display. In order to qualify as radiation the wave should transport energy, that is, propagate. Evanescent fields exist only near the current or charge, see https://en.wikipedia.org/wiki/Evanescent_field. These do not transport energy away from the source and are not radiation. For slowly time varying currents and charges the fields are nearly purely evanescent. The faster the time variation, the higher the fraction of propagating fields is. Regular induction in a coil satisfies Maxwell’s equation (usually the integral form is used) and it’s usually not considered radiation. Yes, you can have changing fields without radiation. I suggest looking at the electric and magnetic fields derived from the L-W equation, which gives different insights from Maxwell's equations. Here is one sample derivation. The electric field of a point charge moving at a constant velocity $$v$$, as derived from the Liénard-Wiechert equation, reads $$E = \frac{q}{4\pi\epsilon_0} \frac{(1-\beta^2)}{(1-n\cdot\beta)^3}\frac{n-\beta}{|r-r_s|^2},$$ where $$r$$ is the location of the stationary target when the force is applied $$r_s$$ is the location of the source when the force leaves it $$n =\overline{r-r_s}$$ and $$\beta = v/c$$. The rest of the electric force equation depends on acceleration, and that part is considered radiation. This part will change when $$v \not= 0$$. This force is correct in a frame where the target does not move. If we choose a frame where source and target both move in a direction unaligned with $$r-r_s$$, the electric force equation fails, and the magnetic force equation then provides a fudge factor which corrects the error. I want to point out that the electric and magnetic fields do not create each other. A moving electric charge creates both of them. The electric charge is the creator, and the fields do not affect each other after they are created. "But are not all changing magnetic or electric fields EM waves?" No, not all. First, charges that move without acceleration can create electric fields that change but that do not include radiation. Second, not all radiation is waves as we usually think of them. Apparently any acceleration of a charge creates radiation. You get a wave if the radiation travels in a periodic way. So if a charge travels up and down in a sine wave, it creates a linearly-polarized wave, polarized up-down. A radio tower does this. Directly above the radio tower the signal is minimized. If a charge travels in a circle, then if you stand edge-on to the circle the radiation is linearly polarized perpendicular to the axis of rotation. Along the axis there is a wave that is circularly polarized. In between there are combinations of both and the sum is elliptically polarized. If a charge travels in some erratic pattern it doesn't exactly make a wave. But you can make a fourier transform to convert whatever it does into some combination of sine waves. In that sense, any charge motion whatsoever is a wave of some sort. Is a static electric or a static magnetic field without charges possible? It is obvious that a static electric field we get by separating electric charges. How about a static magnetic field? We get it by the alignment of the magnetic dipole moments of the involved subatomic particles. This arrangement is static only if the self-inductance is prevailing the thermic movement of the subatomic particles. This is the case for permanent magnets at room temperature or materials very close to zero Kelvin. Trying to find electric or magnetic fields without separated or aligned charges is inconceivable. Conclusion: Static electric and static magnetic fields without charges are impossible. (Mostly the charge is the electron with its electric charge and its magnetic dipole moment.) How do we induce an electric and a magnetic static field? The charge separation as the reason for any electric field is accompanied by an electric current. Faraday induced an current by rotating a conduction disc inside an external magnetic field. The only interaction between the moving electrons and the magnetic field could be the electrons magnetic dipole moment. With the magnetic dipole moment is a spin associated and this together is the reason for the charge deflection and separation. (The inverse process for example are the Lorentz force or the Hall effect.) The purest process is the one, we observe in free space. An electron - visible in a cloud chamber - gets deflected under the influence of an external magnetic field and moves in a spiral path until its kinetic energy is exhausted. The kinetic energy by this gets emitted as EM radiation. Conclusion: A charge separation in an inducting process without the emission of EM radiation is impossible. The same holds for a static magnetic field of aligned magnetic dipoles. The energy for the alignment has to be realized during the disalignment and this again happens in the form of EM radiation. Electromagnetic propagation The only case in which the magnetic and an electric field are independent from charges is the emission of photons. A selfinducting portion of energy is emitted from an accelerated charge and this quanta propagates in free space until it gets absorbed by another subatomic particle. The observation, that a moving charge gain a magnetic field due to its relative motion to the observer, is a conclusion from 200 years ago without the knowledge of the nature of charges: their intrinsic electric field, their intrinsic magnetic dipole and their intrinsic spin with the emission of EM radiation during accelerations.
2022-10-06 17:17:41
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 10, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7596889734268188, "perplexity": 308.4584837788024}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337853.66/warc/CC-MAIN-20221006155805-20221006185805-00233.warc.gz"}
https://mathematica.stackexchange.com/questions/127714/nintegrate-which-includes-a-definite-integral
# NIntegrate Which Includes a Definite Integral I need to calculate a numerical integral which consists of the following functions: f7 := 1/ (Pi^8*( (1/8)*(y0 + y1 + y2 + y3 - y4 - y5 - y6 - y7)^2 + 1) * ((1/8)*(y0 + y1 - y2 - y3 + y4 + y5 - y6 - y7)^2 + 1) * ((1/8)*(y0 - y1 + y2 - y3 + y4 - y5 + y6 - y7)^2 + 1) * ((1/8)*(y0 - y1 - y2 + y3 - y4 + y5 + y6 - y7)^2 + 1) * ((1/8)*(y0 - y1 - y2 + y3 + y4 - y5 - y6 + y7)^2 + 1) * ((1/8)*(y0 - y1 + y2 - y3 - y4 + y5 - y6 + y7)^2 + 1) * ((1/8)*(y0 + y1 - y2 - y3 - y4 - y5 + y6 + y7)^2 + 1) * ((1/8)*(y0 + y1 + y2 + y3 + y4 + y5 + y6 + y7)^2 + 1)) f6 := Integrate[f7, {y7, -Infinity, Infinity}, Assumptions -> {Element[{y0, y1, y2, y3, y4, y5, y6},Reals]}] h7 = NIntegrate[f7*Log[f7/f6], {y0, -Infinity, Infinity}, {y1, -Infinity, Infinity}, {y2, -Infinity, Infinity}, {y3, -Infinity, Infinity}, {y4, -Infinity, Infinity}, {y5, -Infinity, Infinity}, {y6, -Infinity, Infinity}, {y7, -Infinity, Infinity}] My question here is, f6 is hard to calculate symbolically. However, I only need h7. Is there any way to calculate h7 using NIntegrate without trying to calculate f6 symbolically? • People here generally like users to post code as Mathematica code instead of images or TeX, so they can copy-paste it. It makes it convenient for them and more likely you will get someone to help you. You may find this this meta Q&A helpful Oct 1 '16 at 17:01 • Don't use subscripts. Subscript[y, 1] doesn't create a variable $y_1$, it takes a variable y that is inside a two-argument function Subscript. Compute the following to see what I mean: D[Subscript[y, 1], y]. Use simply y1, y2 etc. Oct 1 '16 at 18:07 You can do the multiple Integral completeley with NIntegrate. Try this code: In[31]:= f7[y0_, y1_, y2_, y3_, y4_, y5_, y6_, y7_] = 1/(Pi^8*((1/8)*(y0 + y1 + y2 + y3 - y4 - y5 - y6 - y7)^2 + 1)*((1/8)*(y0 + y1 - y2 - y3 + y4 + y5 - y6 - y7)^2 + 1)*((1/8)*(y0 - y1 + y2 - y3 + y4 - y5 + y6 - y7)^2 + 1)*((1/8)*(y0 - y1 - y2 + y3 - y4 + y5 + y6 - y7)^2 + 1)*((1/8)*(y0 - y1 - y2 + y3 + y4 - y5 - y6 + y7)^2 + 1)*((1/8)*(y0 - y1 + y2 - y3 - y4 + y5 - y6 + y7)^2 + 1)*((1/8)*(y0 + y1 - y2 - y3 - y4 - y5 + y6 + y7)^2 + 1)*((1/8)*(y0 + y1 + y2 + y3 + y4 + y5 + y6 + y7)^2 + 1)); In[32]:= f6[y0_?NumericQ, y1_?NumericQ, y2_?NumericQ, y3_?NumericQ, y4_?NumericQ, y5_?NumericQ, y6_?NumericQ] := NIntegrate[ f7[y0, y1, y2, y3, y4, y5, y6, y7], {y7, -Infinity, Infinity}, MaxRecursion -> 100] (h7 = NIntegrate[ f7[y0, y1, y2, y3, y4, y5, y6, y7]* Log[f7[y0, y1, y2, y3, y4, y5, y6, y7]/ f6[y0, y1, y2, y3, y4, y5, y6]], {y0, -Infinity, Infinity}, {y1, -Infinity, Infinity}, {y2, -Infinity, Infinity}, {y3, -Infinity, Infinity}, {y4, -Infinity, Infinity}, {y5, -Infinity, Infinity}, {y6, -Infinity, Infinity}, {y7, -Infinity, Infinity}, MaxRecursion -> 20]) // Timing NIntegrate::eincr: The global error of the strategy GlobalAdaptive has increased more than 2000 times. The global error is expected to decrease monotonically after a number of integrand evaluations. Suspect one of the following: the working precision is insufficient for the specified precision goal; the integrand is highly oscillatory or it is not a (piecewise) smooth function; or the true value of the integral is 0. Increasing the value of the GlobalAdaptive option MaxErrorIncreases might lead to a convergent numerical integration. NIntegrate obtained -7.47918 and 4.45574566213166 for the integral and error estimates. >> {50349.7, -7.47918} In[33]:= (h7 = NIntegrate[ f7[y0, y1, y2, y3, y4, y5, y6, y7]* Log[f7[y0, y1, y2, y3, y4, y5, y6, y7]/ f6[y0, y1, y2, y3, y4, y5, y6]], {y0, -Infinity, Infinity}, {y1, -Infinity, Infinity}, {y2, -Infinity, Infinity}, {y3, -Infinity, Infinity}, {y4, -Infinity, Infinity}, {y5, -Infinity, Infinity}, {y6, -Infinity, Infinity}, {y7, -Infinity, Infinity}, MaxRecursion -> 100, Method -> {"GlobalAdaptive", "MaxErrorIncreases" -> 10000, Method -> "GaussKronrodRule"}]) // Timing Out[33]= \$Aborted ` Use ?NumericQ for dhe definition of f6 in order to force it to be evaluated only when it has got numerical values from the h7 - Integration. I first got a result with a very large error after one day of calculation on a intel i7 4790 Processor. To get a better result, you should try the code wiht higher MaxErrorIncreases. Attention:This may take a few days of calculation.
2021-09-24 18:07:41
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.379117876291275, "perplexity": 1496.973395635598}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780057564.48/warc/CC-MAIN-20210924171348-20210924201348-00055.warc.gz"}
http://mathonline.wikidot.com/orientable-graphs
Orientable Graphs Definition: A digraph $D$ is said to be Orientable if there exists an orientation (a selection of arc directions) that makes $D$ strongly connected (there exists a directed path between each pair of vertices). Definition: An Orientation of a graph $G$ is the replacement of all edges $\{ x, y \} \in E(G)$ with either the arc $(x, y)$ or $(y, x)$. A good example of an orientable graph is the following digraph $D$: Hence, the original digraph $D$ is orientable.
2020-05-31 14:27:55
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7471122145652771, "perplexity": 213.70161142675073}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-24/segments/1590347413406.70/warc/CC-MAIN-20200531120339-20200531150339-00478.warc.gz"}
https://www.math.ucdavis.edu/research/seminars/?talk_id=3376
# Mathematics Colloquia and Seminars Return to Colloquia & Seminar listing ### Robust Gaussian noise stability Mathematical Physics & Probability Speaker: Joe Neeman, UC Berkeley Location: 1147 MSB Start time: Wed, Nov 7 2012, 4:10PM Given two Gaussian vectors that are positively correlated, what is the probability that they both land in some fixed set A? Borell proved that this probability is maximized (over sets A with a given volume) when A is a half-space. We will give a new and simple proof of this fact, which also gives some stronger results. In particular, we can show that half-spaces uniquely maximize the probability above, and that sets which almost maximize this probability must be close to half-spaces.
2020-01-23 16:32:56
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.96163010597229, "perplexity": 1549.5931391014863}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-05/segments/1579250611127.53/warc/CC-MAIN-20200123160903-20200123185903-00275.warc.gz"}
https://math.stackexchange.com/questions/1531152/functions-of-random-variables
# Functions of random variables Please consider Example 4.8 on http://www.probabilitycourse.com/chapter4/4_1_3_functions_continuous_var.php Would someone mind explaining how $P(X^2≤y)=P(−\sqrt{y}≤X≤\sqrt{y})$ and how $R_y=[0,1]$ (essentially if $-1<x<1$ how does it become $0<y<1$)? Thanks. For your first question, we have $$P\left(X^2\leq y\right)$$ $$=P\left(\sqrt{X^2}\leq \sqrt y\right)$$ $$=P\left(\left|X\right|\leq \sqrt y\right)$$ $$=P\left(−\sqrt{y}\leq X\leq\sqrt{y}\right)$$ For your second question, we have $$-1\lt x\lt 1$$ Which can be rewritten as $$\left|x\right|\lt 1$$ Or $$0\leq\left|x\right|\lt 1$$ Have a look at absolute value. Clearly $\mathsf P(X^2\leq y) = \mathsf P({-}\surd y\leq X\leq \surd y)$ (for any real-valued random-variable $X$, and any non-negative real-value $y$), because if $0\leq X^2\leq y$ , then either $0\leq X\leq \surd y\;$ or $\;{-}\surd y\leq X\leq 0$, and vice versa. Likewise if $-1 < X < 1$ then $0\leq X^2 < 1$; because the square of a negative real-value is a positive real-value, and so is the square of a positive real-value. Hence $X\in (-1;1) \iff X^2\in [0;1)$.
2019-08-18 05:47:28
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9402680397033691, "perplexity": 312.6020986813812}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027313617.6/warc/CC-MAIN-20190818042813-20190818064813-00085.warc.gz"}
https://www.tug.org/pipermail/texhax/2012-February/018937.html
# [texhax] style latex Uwe Ziegenhagen ziegenhagen at gmail.com Tue Feb 7 13:50:48 CET 2012 > On Feb 7, 2012, at 6:14 AM, Mohamed HOUSSNI wrote: > >> Can someone help me ? I need a Style LaTeX to automatically put the first letter of a title in capitals. >> >> Any one have an idea ? > > This is called a drop cap'' for short (long version would be dropped capital''). > > The usual package for doing this is lettrine: > > \documentclass[12pt]{article} > \usepackage[english]{babel} > \usepackage{type1cm, lettrine, blindtext} > \begin{document} > \lettrine[lines=3, findent=3pt, nindent=0pt]{O}{n} the other hand, we denounce with righteous indignation and dislike men who are so beguiled and demoralized by the charms of pleasure of the moment, so blinded by desire, that they cannot foresee the pain and trouble that are bound to ensue; and equal blame belongs to those who fail in their duty through weakness of will, which is the same as saying through shrinking from toil and pain. > \end{document} > > (example modified from http://texblog.org/2011/05/21/drop-caps-with-lettrine/ ) > > William > If you had e.g. to modify the first word in each paragraph, the chickenize package for luaTeX could be helpful. The author, Arno Trautmann, showed really nice examples for luaTeX at the last Dante meeting here in Germany. Uwe
2018-02-19 14:10:25
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9315232038497925, "perplexity": 8248.642190009554}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-09/segments/1518891812665.41/warc/CC-MAIN-20180219131951-20180219151951-00491.warc.gz"}
http://www.math-only-math.com/properties-of-elements-in-sets.html
# Properties of Elements in Sets The following properties of elements in sets are discussed here. If U be the universal set and A, B and C are any three finite sets then; 1. If A and B are any two finite sets then n(A - B) = n(A) – n(A ∩ B) i.e. n(A – B) + n(A ∩ B)  = n(A) 2. If A and B are any two finite sets then n(A ∪ B) = n(A) + n(B) – n(A ∩ B) 3. If A and B are any two finite sets then n(A ∪ B) = n(A) + n(B) ⇔ A, B are disjoint non-void sets. 4. If A and B are any two finite sets then n(A ∆ B) = Number of elements which belongs to exactly one of A or B = n((A – B) ∪ (B – A)) = (A – B) + n(B – A)              [Since (A - B) and (B – A) are disjoint.] = n(A) – n(A ∩ B) + n(B) – n(A ∩ B) = n(A) + n(B) – 2n(A ∩ B) Some more properties of elements in sets using three finite sets: 5. If A, B and C are any three finite sets then n(A ∪ B ∪ C) = n(A) + n(B) + n(C) – n(A ∩ B) – n(B ∩ C) – n(A – C) + n(A ∩ B ∩ C) 6. If A, B and C are any three finite sets then Number of elements in exactly one of the sets A, B, C = n(A) + n(B) + n(C) – 2n(A ∩ B) – 2n(B ∩ C) – 2n(A – C) + 3n(A ∩ B ∩ C) 7. If A, B and C are any three finite sets then Number of elements in exactly two of the sets A, B, C = n(A ∩ B) + n(B ∩ C) + n(C ∩ A) – 3n(A ∩ B ∩ C) 8. If U be the universal set and A and B are any two finite sets then n(A' ∩ B') = n((A ∪ B)') = n(U) -  n(A ∪ B) 9. If U be the universal set and A and B are any two finite sets then n(A' ∪ B') = n((A ∩ B)') = n(U) -  n(A ∩ B) Set Theory Sets
2017-02-21 07:29:18
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8578823208808899, "perplexity": 959.6854407841821}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-09/segments/1487501170696.61/warc/CC-MAIN-20170219104610-00090-ip-10-171-10-108.ec2.internal.warc.gz"}
https://www.physicsforums.com/threads/how-can-i-solve-for-l.749908/
# How can I solve for L? 1. Apr 21, 2014 ### Loopas Hi everyone, For my physics lab I have to solve for inductance, given the following equation and values for C, R, and f: 2*pi*f = $\sqrt{\frac{1}{LC}-\frac{R^{2}}{4L^{2}}}$ I'm just not really sure how to solve for L. I started by squaring both sides but it just seems like I can't solve for inductance because there's two terms for L. Thanks 2. Apr 21, 2014 ### Staff: Mentor After squaring collect everything on the left hand side (leaving zero on the right). Put over a common denominator...
2017-10-23 08:53:57
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5191853642463684, "perplexity": 955.3981184604082}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187825812.89/warc/CC-MAIN-20171023073607-20171023093607-00144.warc.gz"}
https://yro.slashdot.org/story/04/06/10/218206/slashback-nigritude-indignation-artifacts
Please create an account to participate in the Slashdot moderation system typodupeerror DEAL: For $25 - Add A Second Phone Number To Your Smartphone for life! Use promo code SLASHDOT25. Also, Slashdot's Facebook page has a chat bot now. Message it for stories and more. Check out the new SourceForge HTML5 Internet speed test! × ## Slashback: Nigritude, Indignation, Artifacts181 Slashback brings you updates this evening on a handful of previous and ongoing Slashdot stories: read on below for more on how to manipulate Google rankings, what's wrong with Sun's Java Desktop, Claria's plucky response to L.L. Bean's suit, and a fly in the infinite-twin-primes theory. How to not make friends and influence rankings. Ben Michel contributes an update to the search-engine optimization (SEO) contest mentioned last month, the object of which was for contestants to create a site ranked highest by google for a nonsense phrase, "nigritude ultramarine." Michel writes "The first phase of the competition ended last Monday, and the winner was the owner of a forum called nigritude ultramarine--previously known as Merkey.net. According to Brandon Suit, the owner of this forum, the key to his winning strategy was "getting high PR backlinks"--having other websites with high Page Ranks link to him and vice versa. What impact does this have on SEO, and indeed for the rapidly growing search industry in general? The viability of certain underhanded methods in the pursuit of SEO has been clearly reinforced by many of the results of the contest--both Suit and his closest competitor, Philipp Lenssen, posted links in Wiki Sandboxes in order to better their standing. According to Suit, "If you want to manipulate [Google], you can." While search engines certainly have come a long ways from relevance-based searching, it seems that they still have significant changes to make before they can more accurately order results for any given query. The search engines' creators themselves must make countless revisions in their own, perhaps quixotic, quests to create the perfect tools to retrieve relevant data in the vast, ever-expanding realm of the internet." However, not everyone is as matter-of-fact about this method of increasing search-engine visibility; May Kasahara is one of the webmasters and wiki users who isn't. Kasahara writes: "The Search Engine Optimization contest previously mentioned on Slashdot has had a detrimental effect on wiki users and admins (including myself) lately , as the words 'Nigritude Ultramarine' have been showing up in wiki sandboxes across the web. A search on UseModWiki's homepage brought me to this informative entry, which in turn led me to Nigritude Ultramarine and the Wiki Sandbox Effect [note -- mentioned last week on Slashdot] and to these accompanying comments, mostly from very annoyed wiki users." OK, so maybe "infinite" was a strong word. Prof.Phreak writes "Quoting wikipedia: On May 26, 2004, Richard Arenstorf of Vanderbilt University submitted a 38-page proof that there are, in fact, infinitely many twin primes. On June 3, Michel Balazard of Bordeaux reported that Lemma 8 on page 35 is false.[1] As is typical in mathematical proofs, the defect may be correctable or a substitute method may repair or replace the defect. Arenstorf withdrew his proof on June 8, noting "A serious error has been found in the paper, specifically, Lemma 8 is incorrect"." What are these dashed lines all over your sacred cow? reifman writes "Slashdot's link to my article in the Seattle Weekly helped generate 175,000 page views and numerous letters and comments. The article seemed to touch a nerve in the Mac and Linux communities. I've posted a follow up responding to people's letters." Updates from the Academic Affairs Division. zenrandom writes "As Case Western has just recently been reported, we may as well mention the initiative that will be connecting many schools in Ohio. Oarnet, a part of the Ohio Supercomputer Center and The Ohio State University is building a statewide academic and research fiber optic network. Composed of multiple metro-rings and over 1600 miles of fiber." In unrelated college news, Mirell writes "After the FBI previously investigated an open records request filed for the tunnel blueprints at UT, students decide instead to enter via brute force. Hooligans - 1, War Against Terror - 0." The problem with opening Pandora's Box. WC writes "The previous review on JDS2 ended with no successful installation so it wasn't very helpful on what to expect from the Sun distro. This new review has got a working installation but with a slew of new problems: more installation woes, unusable networking, buggy Nautilus and Mozilla window resizing artifacts among others. The author concludes that JDS2 is --effectively-- nothing but JDS 1.1 with the added Sun server software on top, but the desktop part has the same (and more) issues like JDS1 had." Looking innocent is not their strong suit. tbase writes "As reported on News.com.com, Claria, formerly known as Gator, has sued L.L. Bean, charging the retailer with filing a frivolous lawsuit against its advertisers. As covered in a previous Slashdot story, L.L. Bean has filed suit against current and former Claria advertisers for advertising via pop ups over L.L. Bean's site." This discussion has been archived. No new comments can be posted. #### Slashback: Nigritude, Indignation, Artifacts More Login ## Slashback: Nigritude, Indignation, Artifacts Comments Filter: • #### Google results? (Score:5, Funny) <dieter_chen@yahoo.QUOTEcom minus punct> on Thursday June 10, 2004 @08:01PM (#9393557) Well, I know one way to get great results. All you have to do it be Litigious bastards [sco.com] and you're set. • #### Darl is a googler. (Score:5, Funny) by Anonymous Coward on Thursday June 10, 2004 @08:13PM (#9393625) Darl let it be known today in the conference call that he uses google to search the net. Apparently they irony of that, seeing how Darl's position is that ''linux is destroying the foundation of the industry!'', is completely lost on him. • #### Google stronger than Hulk?!? (Score:2) Even the Nigritude Ultramarine Hulk [komar.org] entered the contest at the last minute, and despite trying to SMASH Google, he only ended up #518 ... ;-) Hulk still thinks Slashdot shoulda thrown their hat in the ring, but they declined, saying it wouldn't be proper [slashdot.org] ... • #### Re:Google results? (Score:2, Funny) Such an effort is doomed to be a miserable failure [whitehouse.gov]. • #### Re:Google results? (Score:2) Better than somebody who waffles [johnkerry.com]. • #### Re:Google results? (Score:2) Waffles? What about Bacon and Eggs [johnkerry.com]. • #### OT: Your sig (Score:2) I modded that cambell's ramen post up, not realizing it was plagarism. It was an honest mistake, as I didn't notice the replies. So in conclusion, there's nothing wrong with the moderation system, it's just that most moderators (me included) just don't read all the replies to comments that we moderate. • #### Re:Google results? (Score:2) How many others have a sig like mine? Of course, it doesn't help, since it's a link to Slashdot itself. While we're at it, try Google's I'm Feeling Lucky on miserable failure [google.com] [whitehouse.gov], weapons of mass destruction [google.com] [blueyonder.co.uk], french military victories [google.com] [albinoblacksheep.com], or anti-war peace protestors [google.com] [albinoblacksheep.com]. • #### Re:Google results? (Score:4, Interesting) on Thursday June 10, 2004 @11:31PM (#9394668) Journal Actually, if you think about it, what we want is "SCO" to point to "litigiousbastards.com" as people are more likely to search for "SCO" than "litigious bastards". Join the fight! Visit SCO's website! [litigiousbastards.com] • #### Re:Google results? (Score:3, Funny) SEO? SCO? Whatever, to me it's all just a load of Santorum [spreadingsantorum.com]. • #### Re:Google results? (Score:2) Well, I know one way to get great results. All you have to do it be Litigious bastards and you're set. Hmm, contest for a highly ranked nonsensical sounding website? Slash Dot Dot Org • #### Re:Google results? (Score:2) Another means is to be an unprincipled [adti.net], ignorant [adti.net], mean [adti.net] liar [adti.net] like Ken Brown. (You also may want to read his rant against offshore outsourcing [adti.net].) • #### Lemma? No kidding. (Score:3, Funny) by Anonymous Coward on Thursday June 10, 2004 @08:08PM (#9393596) Arenstorf withdrew his proof on June 8, noting "A serious error has been found in the paper, specifically, Lemma 8 is incorrect". I guess that Lemma turned out to be a real Lemon, eh? *symbol crash* ba-dom-bom Thank you! I'll be here all week! • #### Re:Lemma? No kidding. (Score:2, Funny) by Anonymous Coward Thank you! I'll be here all week! After a joke like that? Not if someone gets you with those symbols first... • #### Re:Lemma? No kidding. (Score:2, Insightful) Some poor grad student isn't gonna be very much liked by his board. They usually have grad students bitch out lemmas like this - but I might be wrong. If I'm right though, some poor schlub is being handed a Master's and shown the door. Piled Higher and Deeper indeed. • #### PageRank. (Score:4, Interesting) <adwulf@[ ]il.com ['gma' in gap]> on Thursday June 10, 2004 @08:10PM (#9393609) Journal Idea of having lots of high-ranking back-links is most certainly an effective one. I used to have a plain old personal homepage, which was ranked as '1' by google, and then I added a link to my site on my page at h2g2 [bbc.co.uk], and watched my rank go up to 4 within a week. Sadly, it didn't last... • #### Re:PageRank. (Score:3, Interesting) And this property is being exploited endlessly. Google needs to find a way to recognize the contribution of a site, rather than just its content and the links. There are tons of sites out there that the sole purpose is to link to other sites and prop up their rankings. • #### Re:PageRank. (Score:3, Interesting) I added my website to my sig. Within a week, I had a shitload of spam attached to my article #1. At least I'm still the first link when you google for "das megabyte." Like I'm sure you always do. I'm also the third link down when you google for "Sorry, ryan." • #### Claria suing another company??? (Score:5, Funny) on Thursday June 10, 2004 @08:12PM (#9393619) Homepage What is this - a game? Half time - change sides.... I wonder what the half time pep talk would have been like in the Claria camp: Now troops, we're going to need to pay for filing motions, and all that other stuff that comes along with suing someone, as well as our defence lawyers. WADDYA MEAN WE SPENT IT ALL ON REBRANDING???....oh that's right (*fights through the fog of denial*)...we were Gator *blushes* • #### Annoyance (Score:1, Insightful) by Anonymous Coward I am a member of the merkey.net forums. While it is (probably) a good thing to have a Slashdot article, it was really annoying to have "nigritude ultramarine" posted everywhere on our forums for two months. All that for an iPod and a monitor too... • #### Use NOINDEX or Robots.txt for Sandbox? (Score:5, Interesting) on Thursday June 10, 2004 @08:13PM (#9393631) The sandbox is a junk area where users can play right? So change the .htaccess to use Follow, NoIndex and google should be blind to them. Or the robots.txt file to block access. Is there some reason you would want the Sandbox indexed? • #### Re:Use NOINDEX or Robots.txt for Sandbox? (Score:1, Insightful) by Anonymous Coward You mean NoFollow, Index, surely? • #### Re:Use NOINDEX or Robots.txt for Sandbox? (Score:2) Maybe I do. I'm not sure. I don't know how PageRank works, but my idea is that if you tell the spider not to index something, but still follow links then it would break the PageRank score of that link. You would be coming from an unindexed page to an indexed page which would be like starting over. Of course I don't know how the ranking really works. • #### Re:Use NOINDEX or Robots.txt for Sandbox? (Score:2) And how does this prevent sandbox pages from being targetted by spambots? If you think telling a spammer his technique doesn't work will stop him, you don't know spammers very well. • #### Re:Use NOINDEX or Robots.txt for Sandbox? (Score:3, Informative) In addition to making the sandbox harder to find, it also means that spamming that sandbox will be less valuable, since the spam link won't enter into the ranking. Since Wiki Sandboxes are for people who want to use that Wiki, having the sandboxes not show up on searches hurts nobody. That is, nobody except for the Wiki admin, who has the initial nuisance of having to reconfigure apache, set up robots.txt, etc. • #### Re:Use NOINDEX or Robots.txt for Sandbox? (Score:2) And how does this prevent sandbox pages from being targetted by spambots? This is not the spam you are looking for... The parent wasn't refering to e-mail spam, but rather google-spam, jerks posting links to their website on your site to prop up their rankings. The idea was to kill the usefulness of the technique by blocking Googlebot from the areas that are vulnerable to this sort of spam. • #### Re:Use NOINDEX or Robots.txt for Sandbox? (Score:2) Unless a Search engine that disobeys robots.txt gets it and along the way Google picks up on THAT listing. • #### Dear Slashdot (Score:5, Funny) on Thursday June 10, 2004 @08:16PM (#9393649) Dear Slashdot, It seems, unfortunately, that comment #9393632, story #110689, is wrong. Simply take v_0=1, r(v):=(1+\cos v)/\sqrt v, \rho(v)=3/\sqrt v, and \phi(v)=v. I imagine that such a mistake has heavy consequences. Sincerely, Letter • #### Mwahahah (Score:4, Insightful) <thebungi@gmail.com> on Thursday June 10, 2004 @08:17PM (#9393657) Homepage OK, so the "Microsoftie" sacred cow follow up quotes this from a Slashdot comment in the story, theorizing that "he couldn't have put it better": "In his lust to dominate the browser market and bring down Netscape, Bill and his cronies decided to give Internet Explorer away for free. They succeeded in undermining Netscape and getting the lion's share of the browser market, but in the process they got an entire generation of users hooked on getting stuff for free. Once users get a taste of free, getting them to pay for stuff becomes difficult or impossible. Why pay for a browser when I can get it for free? Why pay for an operating system when I can get it for free? Why pay for software when I can get it for free? Why pay for music when I can get it for free? Why pay for movies when I can get them for free? In the end, it isn't just Microsoft that's hurt by this." Heh. This, coming from the "teh softwarez must be free-as-in-um-actually-i'm-just-cheap" crowd (which unfortunately makes up the majority of the people who use open source) is absolutely hilarious. In any case, Microsoft has given software away for ages. Suddenly because they gave away IE, the world is on track to become evil purveyors of stolen... things. If that isn't ironic I don't know what is. • #### Re:Mwahahah (Score:2) I'm in no mood to defend or praise M$, but... okay. Any quote which contains ...in the process they got an entire generation of users hooked on getting stuff for free. - as if that attitude was somehow new (cf. the difficulty in launching a successful micropayment or online-currency biz, '95 - present) must've been written by a twentysomething. ("Not that there's anything wrong with that"...) <grrr> • #### Re:Mwahahah (Score:5, Funny) on Thursday June 10, 2004 @08:37PM (#9393766) No, No - This is good If I ever get an RIAA extortion letter I can just explain that I have a long standing habit of getting free stuff from the internet and that complaining at this late date is pointless because it's become the established norm. Kinda of like common law marriages or squatters rights. • #### WHOA!!!! (Score:2) So, under the prevailing Deep Pockets Theory (When suing, go after anyone with serious money, regardless of relevant relationship to the case''), the RIAA should include Microsoft in every one of these lawsuits. That's one time I'd cheer for MS to win. • #### Re:Mwahahah (Score:5, Insightful) on Thursday June 10, 2004 @09:06PM (#9393880) The concept is basically right, but it's misapplied. The public is addicted to free music and movies because they've been getting them for free on TV and radio for decades. THAT's why P2P is not viewed as wrong by the public- "because TV is free anyhow" Suddenly because they gave away IE, the world is on track to become evil purveyors of stolen... things. That's not what was meant at all. The Slashdotter's theory was that consumers addicted to free software would look for... wait for it... Free Software. • #### Re:Mwahahah (Score:2, Insightful) That's not what was meant at all. The Slashdotter's theory was that consumers addicted to free software would look for... wait for it... Free Software. But Free Software doesn't have to be free software! (Thank you, RMS, for that genious naming scheme.) Remember, it's free as in speech, not free as in beer. • #### Re:Mwahahah (Score:2, Funny) It's spelled "genius", genious. :) • #### Re:Mwahahah (Score:4, Insightful) on Thursday June 10, 2004 @10:38PM (#9394387) • #### Re:Mwahahah (Score:2) My original post was just a quicky joke on the estimate of 8 minutes commercials and 52 minutes of show. Your figures are still off. I just did a google search: tv commercials "minutes per hour" [google.com]. Looking at the top ten results and excluding two UK results and one Jordanian result, the 7 US results paint a consistant picture of 15 minutes per hour of actual commercials (plus network promotions and public service announcements on top of that) during primetime. The quantity of actual programming drops to 39 m • #### Re:Mwahahah (Score:2) I have cable TV, and it often angers me that not only do I have to pay for the service, I still have to sit through commercials anyway. (back in the day, when cable TV was in its infancy, there were no commercials on the tube) So, assuming advertising revenue from commercials on TV is consistent over the years - cable TV must be raking in the bucks - and I have no better experience than I had in the 1970s during any given time slice (i.e. number of minutes of programming versus minutes of commercials). I w • #### Re:Mwahahah (Score:2) That's a good point about TV being free. There's a friend of mine who doesn't have a TV, but likes the new Enterprise show. He downloads them from newsgroups to watch them. I don't think there's a problem with that because if he had a TV and VCR, he would have just been recording it instead. The movie thing, however, you usually have to pay to see. (unless it gets shown on TV, but it is usually edited down to fit the TV time) • #### Re:Mwahahah (Score:1) Heh. This, coming from the "teh softwarez must be free-as-in-um-actually-i'm-just-cheap" crowd (which unfortunately makes up the majority of the people who use open source) is absolutely hilarious. Yep. Guilty as charged. If I wasn't so cheap, I'd not only pay SCO \$699 for a runtime licence, I'd post to slashdot telling everyone what a bargain it was. It makes me wonder: if "greed is good" how come "cheap is bad?" Is frugality only a virtue when practiced by CEOs? • #### Re:Mwahahah (Score:3, Insightful) This, coming from the "teh softwarez must be free-as-in-um-actually-i'm-just-cheap" crowd (which unfortunately makes up the majority of the people who use open source) is absolutely hilarious. And you know that "the majority of the people who use open source" are cheap bastards, how? My experience has been that people who use Unix tend to be technically oriented adults who are more aware of ethics, copyrights and patents than the general population. It is self-evident that the Microsoft-using population i • #### Microsoft, OSS, Black (Score:3, Informative) Trust a Winnut to twist this in to some kind of anti-MS / anti-OSS statement. But hey - let's start off by making completely unrelated attacks: Heh. This, coming from the "teh softwarez must be free-as-in-um-actually-i'm-just-cheap" crowd (which unfortunately makes up the majority of the people who use open source) is absolutely hilarious. Hey - let's ignore the Windows-based warez scene. Windows freeware, shareware, and spyware. Let's not bother ourselves with how gleefull Winnuts get when Microso • #### Re:Mwahahah (Score:2) You wrote: Heh. This, coming from the "teh softwarez must be free-as-in-um-actually-i'm-just-cheap" crowd (which unfortunately makes up the majority of the people who use open source) is absolutely hilarious. By calling OSS users cheap, I take it you mean they are less willing to spend money on software than some other crowd. First of all, why is that a bad thing? Secondly, compared to what other crowd? I might point out in response that OSS users are more willing to spend their time learning how to use so • #### JDS (Score:3, Insightful) on Thursday June 10, 2004 @08:20PM (#9393676) kinda interesting to watch the JDS effort. You would think that someone like Sun has nerves,resources,etc. to pull a decent Linux desktop. The reasons it is not are probably combination of: internal apathy of the development group - Linux, desktop, whatever.. Any AC from Sun can comment? cluelessness of the upper management - there is no marketing plan, they just grasping the straws wrong marketing (different from cluelessness). Wtf it is called "Java"? To me, JDS would mean a Swing-based desktop shell on top of very thin Linux distro. Now, that would be innovation. Overall, the JDS just confirms the point that you do not have to be a multi-billion-dollar co to produce major product and when you ARE a m-b-d co, your product may still suck. The innovation is the field owned by talented individuals and hungry startups. • #### Re:JDS (Score:4, Insightful) by Anonymous Coward on Thursday June 10, 2004 @08:39PM (#9393776) Or more likely an incompetant reviewer. I have had JDS2 running on several machines for over a week now and have not experienced a single one of the problems this reviewer mentioned. • #### Re:JDS (Score:3, Insightful) This is Eugenia that we all know and love ;-) • #### Re:JDS (Score:2) Eugenia \Eu*ge"ni*a\ ([-u]*j[=e]"n[i^]*[.a]), n. [NL. Named in honor of Prince Eugene of Savoy.] (Bot.) A genus of myrtaceous plants, mostly of tropical countries, and including several aromatic trees and shrubs, among which are the trees which produce allspice and cloves of commerce. I don't get. • #### Re:JDS (Score:2) I smell an astroturfer. Two reputable reviewers found major problems with JDS2, and some AC on Slashdot negates them? Hmmm... -Jem • #### Re:JDS (Score:3, Interesting) Here's another review [slashdot.org] if you're interested. • #### Re:JDS (Score:2, Insightful) According to Sun, they're capitalizing on the Java® brand recognition... ...which is dumb. We, the folks who know what things like "Linux," "desktop," and "Java" mean from a technical point of view, would likely be confused. I know I was before I read into it. My first reaction was, "Why the hell are they building a Linux desktop distro in Java?" (For the record, my second thought was, "How the hell are they building a Linux distro in Java!!?!") This is a good example of what happens when Marketing • #### Re:JDS (Score:5, Insightful) on Thursday June 10, 2004 @10:23PM (#9394284) I have installed JDS on at least 5 machines at work, mostly low end 2 year old pc's, but also 1 v60x (dual booting suse). JDS seems pretty on target for Sun's stated market, i.e. limited functionality/requirements people. Call centers, hospitals, etc.... Basically, people who need word processing, spreadsheets, email, and and a browser. Sun is not targeting Slashdot readers for JDS users, although a some number us will probably end up administrating them. Sun is not trying to create a sexy, leading distribution, but instead is trying to replace (in their eyes) buggy, insecure, virus ridden, expensive windows machines. Of course, all slashdot readers know this, but still insist on judging the distribution on whether it meets their personal needs, including any wingnut hardware combination they may have. Of course its not a perfect distribution, but that's not Sun's pitch. They are going in to companies saying "Look, we have a product that is more secure, less expensive, and provides all the fuctionality your people need. Plus, we will support it, including phone support." I think the "Java" in JDS is a hint that the linux base is not important to Sun. My guess is that they are trying to get to a Common Desktop Environment ;-) across all their platforms, sparc, x86, and thin client (sun ray). Using the Java name is just a marketing thing now, but it could be more meaningfull (project looking glass, anyone?) in the future. • #### Re:JDS (Score:3, Interesting) Sort of off topic, but JDS is starting to show up in screenshots of linux machines, even when the main focus of the article isn't JDS. See Windows Compatibility for the Linux Desktop [linuxdevcenter.com] for an example. • #### Nonsense Phrase? (Score:1, Funny) the object of which was for contestants to create a site ranked highest by google for a nonsense phrase, "nigritude ultramarine." I was picturing a black supersoldier, kinda like my old drill instructor at Fort Benning. That guy was freakin scary. He was six foot five and hard as a rock in both physique and attitude and had those crazy looking eyes like Bernie Mac. Once out on bivouac he actually opened up a bit and told us about his days in Vietnam when he was a seventeen year old private. • #### My turn to use '9-11'! My turn! My turn! (Score:5, Insightful) <teamhasnoi @ y a h o o .com> on Thursday June 10, 2004 @08:33PM (#9393747) Journal "It's a Homeland Security issue," he said. "We've been aware of the security aspects of the tunnels for a long time. We became more conscious of the security needs after 9-11." Jesus, does everything have to be a Homeland Security issue and tied to 9-11? Whatever happened to harmless breaking and entering? Really, what the hell is the impotent Homeland Security department going to do? Guard the tunnel entrances? Overreact and send the students to Git-mo? Terrorism is old and busted, and is nothing but a political tool and soundbite op. Those who respond, "tell that to the victims of 9-11", I submit that if all those people were here today they would be pretty fucking pissed at all the unconstitutional bullshit that has been done in their name. • #### Re:My turn to use '9-11'! My turn! My turn! (Score:4, Insightful) <cgrrr@nOSPaM.grrr.net> on Thursday June 10, 2004 @08:38PM (#9393773) Homepage Journal ...does everything have to be a Homeland Security issue and tied to 9-11? Only so long as it works... Remember It's for the children! ? <grrr> • #### Re:My turn to use '9-11'! My turn! My turn! (Score:1, Insightful) by Anonymous Coward Whatever happened to harmless breaking and entering? You mean like, what it clearly states the students were charged with at the beggining of the article? Really, what the hell is the impotent Homeland Security department going to do? Nothing, I'd assume, since the article makes no mention of the Homeland Security department, the FBI, the Austin Joint Terrorism Task Force any other government organzation, and that quote was made by the "associate director of utilities and energy management" at the colle • #### Re:My turn to use '9-11'! My turn! My turn! (Score:3, Interesting) "Terrorism is old and busted, and is nothing but a political tool and soundbite op." In a political sense, yes. In a practical sense, not on your life. Just like "the war on drugs" was mostly a political tool, there was a REAL underlying problem that existed (and still exists). Just because some polish up phrases and use them as soundbites doesn't mean the root cause isn't important. There's some very real people out there that want to do some very real harm to our very real civilians. Are they as numer • #### Re:My turn to use '9-11'! My turn! My turn! (Score:4, Interesting) <teamhasnoi @ y a h o o .com> on Thursday June 10, 2004 @11:23PM (#9394617) Journal I'm glad you brought up the 'War on Drugs' - I agree that there are problems with drugs, terrorism, piracy, poverty, illiteracy, and many other things. Why is it that every time that a 'War' is declared, we end up fighting the symptoms of the problems and not the root causes? We'd probably be somewhat safer from terrorists if we'd stop training them and giving them weapons... • #### Re:My turn to use '9-11'! My turn! My turn! (Score:5, Interesting) on Friday June 11, 2004 @03:55AM (#9395777) Homepage There's some very real people out there that want to do some very real harm to our very real civilians. Are they as numerous as people say? No. Are people be incorrectly persecuted in this country for it? Yes. There seems to be a growing trend of people interviewing themselves as a rhetorical technique. Does John Ashcroft do it every time he's on TV? Yes he does. Is it an effective means of controlling the direction of the interview? You bet. Is it starting to grate on my nerves? Absolutely! Will it get old and go away soon? We can only hope. (Yeah, it's off-topic, but when the topic is "Nigritude Ultramarine", so is pretty much everything else) • #### Glad I did this before 911 (Score:3, Interesting) Uni of Queensland [uq.edu.au] has a tunnel network underneath it's central buildings aswell. I am told that it was originally designed to be used by the army in emergencies, but I can't find any record of that (although I believe this is a picture [uq.edu.au] of the tunnel construction). After a few years mapping what entrances were visible, we found a grate that had been left open, so those of us who dared went for a jaunt. They must have had silent alarms aswell (I saw the sensors) so I knew we wouldn't have long. The group • #### Re:Glad I did this before 911 (Score:3, Interesting) Sounds like you might enjoy some recreational urban [infiltration.org] exploration [www.uer.ca]. • #### Re:My turn to use '9-11'! My turn! My turn! (Score:2, Insightful) by Anonymous Coward It's like the Reichstags fire in the Weimar Republic (Germany). Blamed on "them nasty jews" and used as an excuse to round up undesireables and put them into concentration camps. You'll eventually find out that America is not one iota better, and that you have given up just as many rights as the Germans did in 1933. Sure, sure, you are still allowed to vote; but as long as they count the votes what are you going to do about it? • #### Re:My turn to use '9-11'! My turn! My turn! (Score:2, Insightful) Terrorism is old and busted, and is nothing but a political tool and soundbite op. Those who respond, "tell that to the victims of 9-11", I submit that if all those people were here today they would be pretty fucking pissed at all the unconstitutional bullshit that has been done in their name. Agreed. If I died in any manner, and someone used my death to justify fucking up my country, I'd be pretty god damned pissed. (Well, moreso.) What are we doing about it? • #### Re:My turn to use '9-11'! My turn! My turn! (Score:3, Interesting) I agree with you, teamhasnoi. A couple nights ago I watched Ashcroft squirm while Congress grilled him (replayed from that morning on CSpan). Boy that was high entertainment. (He almost got jail time for "Contempt of Congress. ;) But what really caught my attention when he protested that some Patriot Act provision (something to do with police hijinks sans subpoena) was just an extension to terrorism of what they already could do with health insurance fraud. Health insurance fraud! Sound like some congress c • #### For the Wiki Sandboxes (Score:5, Informative) on Thursday June 10, 2004 @08:41PM (#9393789) Homepage Journal Why do they not just disable links to outside pages entirely? It is experimental, right? So why have links to other websites at all? Links could create a bogus page like: You have linked to the URL: http://somesite.wherever.net Internal links within the wiki could be preserved. • #### Re:For the Wiki Sandboxes (Score:2, Insightful) Someone learning might want to test ceating outside links. I think the noindex option is better so that sandboxes are invisible to the google bots. • #### Not so great an idea. Don't follow Phillip. (Score:3, Insightful) Why do they not just disable links to outside pages entirely? It is experimental, right? So why have links to other websites at all? Links could create a bogus page like: you have linked to the URL: http://somesite.wherever.net That's not very helpful because it would be difficult to test the link that way. The idea is to encourage legitimate users to actually make and edit pages. It is unfortunate that Wiki site administrators have to do anything at all. Phillip admits that he does not get it: I • #### Re:Not so great an idea. Don't follow Phillip. (Score:2) It's a sandbox. It's there so that people can experiment there instead of messing up the other sections. Did he use up inordinate amounts of bandwidth? Did he use up tons of CPU? I doubt it - they weren't even complaining about that. If you say "Please test here", heck why complain if people actually do so? Based on the complaints by Wiki owners it's a fact that at either Wiki software is not up to handle these things or Wiki owners aren't. If people bring to attention flaws in code (in this case inadverten • #### Re:Not so great an idea. Don't follow Phillip. (Score:2) Based on the complaints by Wiki owners it's a fact that at either Wiki software is not up to handle these things or Wiki owners aren't. It's the format of wiki that's partially to blame-- most wiki owners want to keep the things open to the public at large. This is assuming that the best way to control the content of a wiki (including the Sandbox) is to regulate its users, and require registration/passwords. The Nigritude spam is a nuisance, one that I shouldn't have to deal with. It's not that I'm not up • #### Re:Not so great an idea. Don't follow Phillip. (Score:2) The wiki stuff is nice and all that. But what if people subtly alter stuff (e.g. history). Would people notice errors? It's just like the patent office problem - they let through almost any crap nowadays that looks half intelligent to the examiner. Would really attacking wikis be worth it? (This nigritude stuff is nothing). Maybe not at the moment so at least some wikis are safe. And maybe just captchas and similar stuff will keep most (not all) wikis safe from automated spamming whilst not requiring the us • #### Re:For the Wiki Sandboxes (Score:5, Insightful) on Thursday June 10, 2004 @10:02PM (#9394180) Homepage I think a better option is a meta tag that lets search engines know that there's user contributed content on the page. (Or maybe something in robots.txt) Google could still index the page, but weigh links on the page lower. -- not a .sig • #### More Googlebombing (Score:5, Interesting) <xptical@@@gmail...com> on Thursday June 10, 2004 @08:52PM (#9393829) Homepage Recently my local computer user group's blog was spammed with user registration. The same user registered about 200 times with slightly different user names and all his home pages linked to the same website. The user never needed to post a single comment in our forums, just the registration page alone gave him 200 links to his home page. If you wanna read a more detailed account of how this works, read here [britecorp.co.uk]. • #### Lemma 8 (Score:1) Lemma 8: If there are a unlimited number of prime pairs implies there are a unlimited number of prime pairs. • #### Wiki Sandboxes exist to TEST! (Score:1) I don't understand what WIKI site owners have against people putting links in sandboxes. The sandboxes are there for users to PLAY IN and test Wiki commands. Nobody is harmed with some silly links in there. The next person in wipes them out (usually). Also most sites clean their sand daily. -- 3D Photography [callipygian.com] • #### Nigritude? (Score:1, Funny) by Anonymous Coward What did you just call me? You racist son of a bitch! I'll fucking kill you and your fucking family. • #### Meh...if I were google i'd have... (Score:4, Funny) <phil@nOspam.webstarsltd.com> on Thursday June 10, 2004 @09:21PM (#9393960) ...made all searches for nigritude ultramarine return results in a random order, shouldn't be that hard for them to do, would have been side splittingly funny as well, hah...SEO spammers...take that, and heres another one for your pointless competition ;o) • #### Re:Randomized Search Results (Score:1) Not hard at all, but I'm glad they didn't. Doing so would have undermined their credibility as being impartial. As I write this, the top link for search engine [google.com] is AltaVista. Google is #7. That's honesty, and they should be proud of it. I really hope they don't become evil after the IPO. • #### Re:Randomized Search Results (Score:2) Poor Yahoo. Their new "Life Engine" ad campaign loses out to "Fractal Life Engine" for link #1. I love it. :) • #### Altavista (Score:4, Interesting) on Friday June 11, 2004 @03:22AM (#9395651) Funny thing is, Altavista ranks itself below Google in a search for search engine. Seems like both search algorithms suffer from low self-esteem. I guess nobody figures on people searching for search engines anyway... • #### So where can we find the old Lemma 8? (Score:2, Interesting) With the author having withdrawn the twin primes paper in the wake of the discovered flaw, arXiv no longer has the original up so we can see what went wrong. Does anyone have a mirror? • #### Re:So where can we find the old Lemma 8? (Score:2, Informative) Withdrawing from arxiv.org doesn't mean the original version goes away: math/0405509v1 [arxiv.org] • #### Re:So where can we find the old Lemma 8? (Score:1) Mea culpa, I saw the '0kb source' in the updated version and assumed that meant it was no longer available. Mod parent down for idiocy on my part, please. :-) • #### PlayingWith Fire (Score:2, Interesting) by Anonymous Coward Nigritude, ouch... Playing with fire. Expect the wrath of the NAACP anyday now. Remember the guy who actually got fired for using the word niggardly? • #### Real Security, re: tunnels (Score:2) UTPD officials say the current security system is adequate. "I think we've got a good system in place, and I think the apprehension of six individuals shows how the system works," Stalder said. So, they caught the perps. That's fine for simple vandalism, but if they continue to hide behind "homeland security," I would demand that they actualy provide that level of security. Specifically, the system failed in three ways: - They don't know if all intrusions resulted in capture. I suspect not; it's the crimina • #### Well lookie there. (Score:2) Wiki admins are irritated that a widespread net meme is showing up in their sandboxes? Heh. All your nigritude are ultramarine to us #### Related LinksTop of the: day, week, month. "I'm not a god, I was misquoted." -- Lister, Red Dwarf Working...
2017-04-30 03:32:14
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.32440316677093506, "perplexity": 6664.040859433597}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917124297.82/warc/CC-MAIN-20170423031204-00204-ip-10-145-167-34.ec2.internal.warc.gz"}
https://codereview.stackexchange.com/questions/83135/designing-a-coffee-machine/83137
# Designing a coffee machine I was recently rejected from what looked like a really promising string of interviews. I did very well in a questionnaire style review, and then they handed me this assignment (more or less): Design a coffee machine which makes different beverages based on set ingredients. The initialization of the recipes for each drink should be hard-coded, although it should be relatively easy to add new drinks. The machine should display the ingredient stock (+cost) and menu upon startup, and after every piece of valid user input. Drink cost is determined by the combination of ingredients. For example, Coffee is 3 units of coffee (75 cents per), 1 unit of sugar (25 cents per), 1 unit of cream (25 cents per). Ingredients and Menu items should be printed in alphabetical order. If the drink is out of stock, it should print accordingly. If the drink is in stock, it should print "Dispensing: ". To select a drink, the user should input a relevant number. If they submit "r" or "R" the ingredients should restock, and "q" or "Q" should quit. Blank lines should be ignored, and invalid input should print an invalid input message. They supplied the default ingredients (&stock @10) and drinks/recipes. They told me to "not try and impress", don't overly document (specifically commenting is not required). The code should be scaleable, readable, and just not overly complicated. Anyway, it's more likely that I made some major error in data structure choice or handling drink selections, than me missing some detail. Still, I can privately send the full assignment description if anyone is curious...I just don't want to put it out there since it's a big company currently hiring. I changed the name of some stuff in my code to make the exact assignment less searchable. I actually have a lot of opinions on things I would change in this code, but I have a hard time believing that they are things that would get me rejected. It must be something bigger... (I have since been informed that the team was hoping for "A more OO design"...felt like I took it in that direction, personally.) DrinkMachine import java.io.BufferedReader; import java.io.IOException; import java.util.Collections; import java.util.List; import java.util.ArrayList; import java.util.Map; public class DrinkMachine{ private static List<Drink> drinkList = new ArrayList<Drink>(); private static List<Ingredient> ingredientList = new ArrayList<Ingredient>(); public static void main(String[] args) { updateCosts(); updateMakeable(); display(); startIO(); } public static void startIO(){ String input = ""; //running loop while(true){ try { /*I considered using a switch here for more readable code. Elected to have a dynamic drink menu so drinks can be added * to addAllDrinks() and this loop would not have to be change. The decision to have slightly less readable code * in favor of a dynamic menu was made upon my belief that easily modifiable code is incredibly important. */ if(input.equals("")){ continue; }else if (input.equals("q")){ System.exit(0); }else if(input.equals("r")){ restockIngredients(); updateMakeable(); }else if(Integer.parseInt(input) > 0 && Integer.parseInt(input) <= drinkList.size()){ //dynamic drink menu selection makeDrink(drinkList.get(Integer.parseInt(input)-1)); }else{ throw new IOException();//legal, but invalid input } } catch (Exception e) { System.out.println("Invalid selection: " + input + "\n");//illegal input } }//running loop } public static void display(){ //The sample IO in the assignment appears to add a blank line after output, so I included that. System.out.println("Inventory:\n"); for (Ingredient i : ingredientList){ System.out.println(i.getName() + "," + i.getStock() + "\n"); } int count = 1; for (Drink d : drinkList){ System.out.printf("%d,%s,$%.2f," + d.getMakeable() + "\n\n", count, d.getName(), d.getCost()); count++; } } public static void updateMakeable(){ for (Drink d : drinkList){ Map<String, Integer> currRecipe = d.getRecipe(); for (Ingredient i : ingredientList){ if (currRecipe.containsKey(i.getName()) && i.getStock() < currRecipe.get(i.getName())){ d.setMakeable(false); break;//check next drink } d.setMakeable(true); }//Ingredient loop }//Drink loop } public static void updateCosts(){ for (Drink d : drinkList){ double currCost = 0; Map<String, Integer> currRecipe = d.getRecipe(); for (Ingredient i : ingredientList){ if (currRecipe.containsKey(i.getName())){ currCost += i.getCost()*currRecipe.get(i.getName()); } }//inner d.setCost(currCost); }//outer } public static void makeDrink(Drink drink){ if(drink.getMakeable()){ System.out.println("Dispensing: " + drink.getName() + "\n"); for (Ingredient i : ingredientList){ if(drink.getRecipe().containsKey(i.getName())){ i.setStock(i.getStock()-drink.getRecipe().get(i.getName())); } } }else{ System.out.println("Out of stock: " + drink.getName() + "\n"); } updateMakeable(); display(); } public static void restockIngredients(){ for(Ingredient i : ingredientList){ i.setStock(10); } updateMakeable(); display(); } //Add ingredients through addAllIngredients public static void addIngredient(Ingredient ingredient){ ingredientList.add(ingredient); } //Add drinks through addAllDrinks public static void addDrink(String name, String[] recipe){ drinkList.add(new Drink(name, recipe)); } public static void addAllIngredients(){ addIngredient(new Ingredient("Coffee", 0.75)); addIngredient(new Ingredient("Decaf Coffee", 0.75)); addIngredient(new Ingredient("Sugar", 0.25)); addIngredient(new Ingredient("Cream", 0.25)); addIngredient(new Ingredient("Steamed Milk", 0.35)); addIngredient(new Ingredient("Foamed Milk", 0.35)); addIngredient(new Ingredient("Espresso", 1.10)); addIngredient(new Ingredient("Cocoa", 0.90)); addIngredient(new Ingredient("Whipped Cream", 1.00)); Collections.sort(ingredientList); } public static void addAllDrinks(){ addDrink("Coffee", new String[]{"Coffee", "Coffee", "Coffee", "Sugar", "Cream"}); addDrink("Decaf Coffee", new String[]{"Decaf Coffee", "Decaf Coffee", "Decaf Coffee", "Sugar", "Cream"}); addDrink("Caffe Latte", new String[]{"Espresso", "Espresso", "Steamed Milk"}); addDrink("Caffe Americano", new String[]{"Espresso", "Espresso", "Espresso"}); addDrink("Caffe Mocha", new String[]{"Espresso", "Cocoa", "Steamed Milk", "Whipped Cream"}); addDrink("Cappuccino", new String[]{"Espresso", "Espresso", "Steamed Milk", "Foamed Milk"}); Collections.sort(drinkList); } } Drink import java.util.HashMap; import java.util.Map; public class Drink implements Comparable<Drink>{ private Map<String, Integer> recipe = new HashMap<String, Integer>();//map ingredients to units per private String name; private double totalCost = 0; private boolean makeable = false; public Drink(String name, String[] recipe){ this.name = name; setRecipe(recipe); } public int compareTo(Drink drink){ return name.compareTo(drink.getName()); } public void setRecipe(String[] recipe){ for(String s : recipe){ if(this.recipe.containsKey(s)){ this.recipe.put(s, this.recipe.get(s)+1);//increment if multiple units }else{ this.recipe.put(s, 1);//insert first occurrence of ingredient } } } public void setName(String name){ this.name = name; } public void setCost(double totalCost){ this.totalCost = totalCost; } public void setMakeable(boolean makeable){ this.makeable = makeable; } public Map<String, Integer> getRecipe(){ return recipe; } public double getCost(){ return totalCost; } public String getName(){ return name; } public boolean getMakeable(){ return makeable; } } Ingredient public class Ingredient implements Comparable<Ingredient>{ private String name = ""; private double cost = 0.00; private int stock = 0; public Ingredient(String name, double cost){ this.name = name; this.cost = cost; this.stock = 10; } public int compareTo(Ingredient ingredient) { return name.compareTo(ingredient.getName()); } public void setName(String name){ this.name = name; } public void setCost(double cost){ this.cost = cost; } public void setStock(int stock){ this.stock = stock; } public String getName(){ return name; } public double getCost(){ return cost; } public int getStock(){ return stock; } } • Without even looking at the implementation if I was reviewing this code for an interview the all the static members and static void methods would really scare me. Most static states (mostly mutable ones) make testing hard and hurts extensibility. – pllee Mar 4 '15 at 20:04 • Also, why are the addAll* methods public? They're only used once at initialization. As a general advice: OOP dictates that you need to know as little as possible about the implementation of a class. There should not be a public update* method (if at all), since one should expect an object of a class to behave properly and figure out by itself when it needs updating. It seems that you're seeing OOP as 'let's split my program into useful subroutines' instead of 'I want to use my classes without knowing their exact implementation'. – Sanchises Mar 4 '15 at 22:49 ## 5 Answers I have since been informed that the team was hoping for "A more OO design"...felt like I took it in that direction, personally. All the methods in your DrinkMachine class are static. I believe the ability to have more than one DrinkMachine is a very important feature. DrinkMachines should be able to have different inventories, different products, etc. Now, you might think "But I was told to keep it simple!". This is not about "trying to impress" or anything, this is about reasonable functionality. I am afraid that this staticness of the DrinkMachine cost you a lot of "interview-points". Do you really need to be able to change the name of an Ingredient after it has been created? I think you should strive on making some of your fields final. This also helps with tim's point about you don't need to initialize fields if you assign them in a constructor, that's just confusing. - final fields can only be initialized once. • I think you're right about how the ability to have more than one DrinkMachine is an important feature, even if it is not a specified requirement. Regarding making some fields final...I suppose I need to think harder about whether or not fields should change. I think I often say "Why not" or "Doesn't hurt" regarding that...in other words, I decided not to make them final because I couldn't think of a reason why they shouldn't be able to change. Instead, I suppose it's better to think "Is there a reason why it ever should change?" and declare it final if I can't think of one. – spacecadet Mar 4 '15 at 8:57 • @spacecadet : this subject is covered by keyword like mutable/immutable. As too many programmers, you have been teached to build automatic mutable objects, which is sad. Try to learn about Immutable advantages : programmers.stackexchange.com/questions/151733/… – Cyril Gandon Mar 5 '15 at 13:24 • @CyrilGandon This is tricky because I know I overdid it by making everything changeable, but when it comes to mutability in general, a lot of my feelings are shared w/ this: programmers.stackexchange.com/a/151786/170331 I ask "Why shouldn't it change?" because mutability is default. If the question should instead be "Why should it change?" then we would want things to be final and immutable by default. Anyway I'm not the one to decide if this school of thought is right/wrong, but it's why I (and apparently others) think mutability until there is an explicit reason for it not to be. – spacecadet Mar 5 '15 at 23:46 First of all, I'm sorry that you didn't get the job. Better luck next time :) Overall, I don't think your code is bad. But I do have a couple of points: OOP • My main complaint would be that you are missing a recipe class. Having it as a string array/map is not very flexible. Because of this, you also expose the inner workings of the drink class to the outside world (for example in updateMakeable). • setRecipe is a bit confusing. Generally, setters only set things, but this method does more. It also doesn't fit in with getRecipe which returns a different type than setRecipe sets. • Your DrinkMachine does too much. It displays, it reads input, it handles the state and logic of the machine, and it contains the test cases and the main loop. At least the first two and the last two case should be handled somewhere else. Misc • you don't need to initialize fields if you assign them in a constructor, that's just confusing. • having the stock of ingredients hardcoded in the constructor doesn't seem very flexible. • add @Override if appropriate (may avoid bugs, and increases readability). • having comments which show where a loop ends is not very common in Java (if your code is so nested that it needs them, it should be refactored). • many of your comments are actually not needed, or should be better integrated into the code (eg legal, but invalid input shouldn't be a comment, but the message of the exception, //running loop isn't very helpful, etc). • short variable names lead to harder to read code (drink, ingredient, currentCost are all not too long, and a lot more readable than d, i, and currCost). • Thanks, lots of good stuff here. A recipe class seems like the big thing. I might have taken some of their tips too literally (simplicity, don't try and impress, etc). I can recall thinking about splitting the code up more, but telling myself that it would go against the simplicity to have too much code in too many different places....I guess that wound up having the opposite effect. Thanks again. – spacecadet Mar 3 '15 at 22:43 • I don't see a need to introduce a Recipe class. After all, what is a Drink, but a name and an ingredient list? The problem is that Drink shouldn't carry coffee machine state such as makability, and Ingredient shouldn't carry coffee machine state such as inventory level. – 200_success Mar 4 '15 at 12:07 I agree with their assessment that the OOP design could be improved. In particular, the way you initialize the data is cumbersome, and I object to the necessity of calling updateMakeable() throughout your DrinkMachine code. ### Ingredient.java What makes an ingredient comparable to another? Is the comparison by name, cost, or quantity? Having the stock be a part of an Ingredient doesn't feel right, as it mixes in part of the DrinkMachine's state. In my opinion, an Ingredient should be stateless and immutable. setName() makes no sense. setCost() is probably over-engineering for this exercise. Since the exercise suggests that hard-coding the ingredients is OK, I would just use an enum whose entries are in alphabetical order. import java.math.BigDecimal; public enum Ingredient { // An extensible design wouldn't use an enum, but the problem specification // allows such hard-coding. These entries must be in alphabetical order. COCOA("0.90"), COFFEE("0.75"), CREAM("0.25"), DECAF_COFFEE("0.75"), ESPRESSO("1.10"), FOAMED_MILK("0.35"), STEAMED_MILK("0.35"), SUGAR("0.25"), WHIPPED_CREAM("1.00"); private final String name; private final BigDecimal cost; Ingredient(String cost) { this.name = this.name().replace("_", " ").toLowerCase(); this.cost = new BigDecimal(cost); } // Using fixed-point representation to prevent rounding problems public BigDecimal getCost() { return this.cost; } public String toString() { return this.name; } } ### Drink.java Many of the same remarks for Ingredient apply to Drink as well. Drinks aren't inherently Comparable. You don't have to plan for product reformulation (setRecipe()), rebranding (setName()) or adjusting the profit margins (setCost()). Makability is part of the DrinkMachine's state. (If you were to have such a method, isMakable() would be a more appropriate name than getMakeable(), both because of the spelling and because of the Java naming convention for predicates.) In getRecipe(), it would be a good idea to return an unmodifiable map. Once again, I would go with an enum, which takes care of initialization and sorting. import java.math.BigDecimal; import java.util.Collections; import java.util.HashMap; import java.util.Map; public enum Drink { // An extensible design wouldn't use an enum, but the problem specification // allows such hard-coding. These entries must be in alphabetical order. CAFFE_AMERICANO("Caffè Americano", "3 ESPRESSO"), CAFFE_LATTE("Caffè Latte", "2 ESPRESSO", "STEAMED_MILK"), CAFFE_MOCHA("Caffè Mocha", "ESPRESSO", "COCOA", "STEAMED_MILK", "WHIPPED_CREAM"), CAPPUCCINO("Cappuccino", "2 ESPRESSO", "STEAMED_MILK", "FOAMED_MILK"), COFFEE("Coffee", "3 COFFEE", "SUGAR", "CREAM"), DECAF_COFFEE("Decaf Coffee", "3 DECAF_COFFEE", "SUGAR", "CREAM"); private final String name; private final Map<Ingredient, Integer> ingredients; private final BigDecimal cost; Drink(String name, String... recipe) { Map<Ingredient, Integer> map = new HashMap<Ingredient, Integer>(); BigDecimal cost = BigDecimal.ZERO; for (String spec : recipe) { String[] amountOfStuff = spec.split(" ", 2); int quantity = (amountOfStuff.length > 1) ? Integer.parseInt(amountOfStuff[0]) : 1; String stuff = amountOfStuff[amountOfStuff.length - 1]; Ingredient ingredient = Enum.valueOf(Ingredient.class, stuff); map.put(ingredient, quantity); cost = cost.add(ingredient.getCost().multiply(new BigDecimal(quantity))); } this.name = name; this.ingredients = Collections.unmodifiableMap(map); this.cost = cost; } public Map<Ingredient, Integer> getRecipe() { return this.ingredients; } public BigDecimal getPrice() { return this.cost; } public String toString() { return this.name; } } ### DrinkMachine.java The improvements above will pay off here: • addAllIngredients() and addAllDrinks() are taken care of by the enums. • updateCosts() is taken care of by the Drink constructor. • updateMakeable() is no longer needed, since makability is no longer a property of each drink. Rather, the machine can query itself to see whether each drink can be made. (See canMake().) Eliminating updateMakeable() is a good thing — needing to call it everywhere is error-prone. Note that in your handling of the "r" command, you call restockIngredients(); followed by updateMakeable(); but restockIngredients() itself already calls updateMakeable(). The DrinkMachine could benefit from having a toString() method. It can take over your display() method. Scanner provides a slightly nicer interface than BufferedReader. In this case, since all you need is readLine(), Console also works nicely. Now that initialization has already been taken care of, your event loop can just be in main(). In the loop, calling System.exit(0) is unjustified — just a break would suffice. Catching all Exceptions is also too strong. For example, end-of-file (Ctrl z on Windows, Ctrl d elsewhere) would fail to cause the program to exit the way I expect. I also find it odd that you classify legal but invalid input as an IOException rather than IllegalArgumentException. import java.io.Console; import java.util.*; public class DrinkMachine { public static class OutOfStockException extends RuntimeException { OutOfStockException(Drink drink) { super(drink.toString()); } } private Map<Ingredient, Integer> stock = new HashMap<Ingredient, Integer>(); public int getStock(Ingredient ingredient) { return this.stock.containsKey(ingredient) ? this.stock.get(ingredient) : 0; } public void restock() { for (Ingredient ingredient : Ingredient.values()) { if (this.getStock(ingredient) < 10) { this.stock.put(ingredient, 10); } } } public Drink[] menu() { return Drink.values(); } public boolean canMake(Drink drink) { for (Map.Entry<Ingredient, Integer> quantStuff : drink.getRecipe().entrySet()) { if (this.getStock(quantStuff.getKey()) < quantStuff.getValue()) { return false; } } return true; } public void make(Drink drink) throws OutOfStockException { if (!this.canMake(drink)) { throw new OutOfStockException(drink); } for (Map.Entry<Ingredient, Integer> quantStuff : drink.getRecipe().entrySet()) { int quant = quantStuff.getValue(); Ingredient stuff = quantStuff.getKey(); this.stock.put(stuff, this.getStock(stuff) - quant); } } public String toString() { StringBuilder out = new StringBuilder("Inventory:\n"); for (Ingredient ingredient : Ingredient.values()) { out.append(ingredient).append(": ") .append(this.getStock(ingredient)).append('\n'); } out.append("\nMenu:\n"); int i = 0; for (Drink drink : Drink.values()) { out.append(++i).append(": ").append(drink); if (this.canMake(drink)) { out.append(",$").append(drink.getPrice()); } else { out.append(" (out of stock)"); } out.append('\n'); } return out.toString(); } private static String prompt(Console con, String prompt) { con.printf("%s", prompt); } public static void main(String[] args) { DrinkMachine machine = new DrinkMachine(); machine.restock(); Console con = System.console(); con.printf("%s\n", machine); String origCmd; while ((origCmd = prompt(con, "Select: ")) != null) { String cmd = origCmd.toLowerCase(); if (cmd.isEmpty()) { continue; } else if ("q".equals(cmd)) { break; } else if ("r".equals(cmd)) { machine.restock(); } else try { int selection = Integer.parseInt(cmd); Drink order = machine.menu()[selection - 1]; machine.make(order); con.printf("Dispensing: %s\n", order); } catch (ArrayIndexOutOfBoundsException | NumberFormatException invalid) { con.printf("Invalid input: %s\n", origCmd); continue; } catch (OutOfStockException outOfStock) { con.printf("Out of stock: %s\n", outOfStock.getMessage()); continue; } con.printf("%s\n", machine); } } } • I really like your enum-approach and I find your code very readable overall (+1). Just one quick question: Is there a reason you are using a BigDecimal for the price rather than a simple double or float or even an int in cents if you're worried about rounding? Reason I'm asking is because @skiwi copied your idea codereview.stackexchange.com/questions/83195/… and it might be helpful to justify or question the use of a very involved object like a BigDecimal for a very simple purpose like this. – Markus A. Mar 4 '15 at 21:44 • @MarkusA. This application is no performance-sensitive. BigDecimal results in simpler code than using an int number of cents. float and double are unable to represent hundredths exactly. BigDecimal is immune to rounding errors. They carry precision information, so that new BigDecimal("1.00") will naturally print as "1.00" rather than "1". Unlike strings, you can do arithmetic on them, which is necessary to arrive at the total cost. For all those reasons, I chose BigDecimal. – 200_success Mar 4 '15 at 23:03 • I didn't know that they preserve their precision when printing. That's indeed helpful. The imprecision that float and double introduce into the numbers, though, are so extremely tiny that you'd have to order tens of thousands of drinks on the same check before you see it for floats. For doubles it's many trillions. Should be safe. :) – Markus A. Mar 5 '15 at 3:59 I believe your interviewer was looking for the usage of the decorator pattern. Atleast, I've seen almost the exact same problem description in a design pattern book I read. It's a very good read and they deserve credit! The decorator pattern allows you to create compositions dynamically, starting from a normal coffee, and adding decoration layers (milk, sugar, etc.) on top of it. Calling the cost function on the highest layer, will then calculate the cost of the highest layer and calls the cost function of the layer underneath, adding both together. A short sketch: base = new Drink(); DrinkInterface topping1 = new Topping(base); DrinkInterface topping2 = new Topping(topping1); I believe it would be pointless to give you the full solution, retrying it with the pattern would be a great exercise! For a more in depth explanation of the decorator pattern (coffeemaker scenario), take a look at this. • Personally, I think avoiding this pattern is what the interviewers meant with not try and impress. – Simon Forsberg Mar 4 '15 at 13:15 • I just don't see the reason for it. Keeping a list of all ingredients seems simpler to me. Using the pattern that you suggest will end up with quite many classes in this case. – Simon Forsberg Mar 4 '15 at 15:20 • Keeping logic separated? Since one should develop with the idea that additional logic is needed. If you want to add a new decorator, all you have to do is create a class that implements the DrinkInterface and it scales perfectly, without having to alter any other existing classes (except the client ofcourse). Also Imagine if they in the future want to add a make() method, which can be different for each topping (milk/sugar/mocca). I'm not saying the pattern is a necessity or golden bullet. I however do believe it's a good idea to make it future proof and it's easier for communication purposes. – DJanssens Mar 4 '15 at 15:35 • One small reason I would shy away from this approach is that the decorator pattern implies an order: it doesn't matter if you add milk then sugar vs sugar then milk. It's also more cumbersome to interact with the drink directly, given that it's buried under layers of reference. – skeggse Mar 4 '15 at 22:55 • Yes - As DJanssens suggested its the Decorator pattern that is the key to this problem...Funnily enough Wikipedia as an example "Coffee Making Scenario" that might have actually been the inspiration for the assignment :-) : See en.wikipedia.org/wiki/… – AdamW Mar 5 '15 at 12:36 Sorry to hear that you didn't get the job. I'll keep my fingers crossed for you for your next attempts. It would probably be very hard to pin-point one exact detail that could have saved this code. It is very possible that their response that they were hoping for a "more OO design" was just some vague thing that they came up with to justify their decision. Since it is very hard to know how other candidates performed in the other parts of the interview, it is not even certain that this code was the main reason that they chose someone else. As these coding exercises are a way for the company to gauge your familiarity and experience with a specific language or with coding in general, it was probably a whole collection of little things that made you seem less experienced than a competing applicant. @Tim already compiled a good list of these sorts of things. But if they were specifically looking for a Java developer, let me add some more observations (not all of which everyone would necessarily care about or even agree with): • You use ArrayLists for your drink and ingredient lists. Since each of these lists should not support two identical entries, the more sensible choice would be to use a Set instead of a List. Specifically, using a TreeSet will always keep its entries sorted, so you would not even need to use Collections.sort(...) any more. Someone might assume that you are not familiar with these "advanced" data-structure types if you are not using them. • Since you are using the String input only inside of the try {}-block, you can just define it right there (String input = reader...) rather than defining it outside and initializing it to empty String even though that initialized value is never used. • You throw an IOException for a wrong input, which is usually understood to relate to I/O errors when it comes to things like reading/writing files or network streams. A more sensible exception would be UnsupportedOperationException or IllegalArgumentException. In general, though, it is extremely expensive to throw and catch exceptions. So the much better approach would be to move your "Invalid selection"-println into the else-block instead of throwing an exception. • It is generally considered bad practice to catch generic Exceptions rather than specific ones, especially if you then don't output the exception! For example, if someone modifies your makeDrink(...)-code and makes a mistake that leads to a NullPointerException or something else being thrown in that code, you catch this exception and turn it into the "Invalid selection"-message, which will make it much harder to track down that bug. • If a for-statement or if-statement only has a single line following it, you can omit the curly braces (e.g. for(Ingredient i : ingredientList) i.setStock(10);). • Java support varargs which allow you to write public static void addDrink(String name, String... recipe) and have it behave exactly as String[] would inside the function, except that you can now call it as addDrink("Coffee", "Coffee", "Coffee", "Coffee", "Sugar", "Cream"). While this seems less awkward than having to add new String[]{}, it does make the code a little less clear as the first "Coffee" in the list gets special treatment but looks like part of the list in the call. So this one is definitely a judgement call. • ... There are probably several other things that they (or others) might consider indications of mastery of a language (or lack thereof), like using the ternary operation and enums (as @200_success did in his answer), maybe even generics if you can find a way to have it make sense in this simple example, ... So, even though they specifically said to "not try and impress", there are definitely certain things that people who live and breathe Java to the point that they are dreaming about it at night would just instinctively do or use to write concise and clear code without the intention to "impress". What exactly those things are depends a lot on the individual person (you can see hints of this in the comments to answers here, although there are definitely some conventions), so it's hard to say what rubbed them the wrong way for your submission (or at least a better way for someone else's). With that in mind, while it will be useful for you to see examples of how this code would have been written by others (@200_success's is a great one!), the only way to truly improve your chances for the same kind of interview question next time is to code, code, code! Or for general inspiration or if you are short on ideas for own projects, you could read over and try to understand code written by others for renowned open-source projects. Large parts of OpenOffice are written in Java, for example. Do realize, though, that not all code written for any of these projects would be considered a stellar example of how to do it right. But it should still give you a great way to find out about language constructs you maybe didn't know about (a ? b : c, for example) and other available data structures. So you can make yourself a lot more resourceful simply by trying to understand every last (content and syntactical) detail of what you read there... Good luck! • I generally agree with your answer here, especially the TreeSet<T> suggestion (I haven't considered that in my own attempt codereview.stackexchange.com/questions/83195/…). I do not agree though with that it is better to omit curly braces, I would say: Never omit them, they can cause nasty bugs. – skiwi Mar 5 '15 at 9:54 • @skiwi True. Leaving them out can make the code more bug-prone. The reasen I added it to the list is that to me, using them in such simple contexts instead of the one-liner I put looks like awkward CS-class-style stick-to-the-rules coding rather than years of in-field experience. But this is a highly subjective interpretation and as far as good coding style goes, you are certainly correct. To me, it's kinda like using "kinda" when writing English. You wouldn't do that if you'd only learned it in school because it's not "correct", but if you use English daily, eventually you might. – Markus A. Mar 5 '15 at 15:21 • Thanks @MarkusA. , I've taken a lot of notes based on your suggestions. A few questions/remarks.. re: dropping curly braces. I find myself often wanting to, but electing to keep them for the sake of readability and, you guessed it, how my professors taught me to keep it. A set / treeset specifically is something that actually clicked once I had submitted it. I'm familiar with a lot of these data-structures...I just don't use them enough. Need more diverse projects! Is there any set context where varargs are considered best practice? Or just personal pref? How about break & continue? – spacecadet Mar 5 '15 at 23:16 • There's nothing wrong with using break and continue. Using break inside the updateMakeable() loop was correct. Changing it to if-else, without a break, would be a logic bug. – 200_success Mar 7 '15 at 8:29 • @spacecadet Actually, I made a mistake: I read your break as a continue (rookie-mistake, I guess ;). I removed that point from my answer. I guess what caught my attention in your updateMakeable()-method was the fact that you repeatedly call d.setMakeable(true) inside the inner for-loop and I mistakenly blamed it onto your use of break. I would move that statement before the ingredients-loop so it only gets called once. – Markus A. Mar 7 '15 at 16:04
2020-07-04 22:40:06
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.23884475231170654, "perplexity": 1683.0470037046032}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-29/segments/1593655886706.29/warc/CC-MAIN-20200704201650-20200704231650-00089.warc.gz"}
https://www.physicsforums.com/threads/how-to-derive-x-t-equation-from-energy.466266/
# How to derive x(t) equation from energy 1. Using V(x)= -max, in the following equation: $$\int_{x_0}^x \frac{dx}{\pm \sqrt{{\frac{2}{m}\{E-V\left( x\right)\}}}} \$$ = t - t0 to get: x = x0 + v0 + at2/2 E is total energy and V(x) is potential energy. I have tried hard integrating it in various ways but do not seem to get the required result. I would really appreciate in help or tips in this regard. When I use E - 0.5mv^2= V(x), the denominator becomes v and really does not help at all. If I do not do that, and use V(x) = -max that does not help either. I do not seem to be reaching the required equation in any way. Last edited: tiny-tim Homework Helper welcome to pf! hi cream3.14159! welcome to pf! (try using the X2 icon just above the Reply box ) … to get: x = x0 + v0 + at2/2 When I use E - 0.5mv^2= V(x) (it should of course be x = x0 + v0t + at2/2) why are you using E - 0.5mv2 ? this is a perfectly ordinary integral of (constant - 2ax)-1/2 show us what you get Hi! Thank you for the response. I solved it with someone's help. The mistake I was doing was to use 0.5mv2-m*a*x to replace E. However, using v0 and x0 instead of v and x in this expression works to give the desired result and also, one has to put t0 = 0.
2021-05-18 12:07:42
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.540822446346283, "perplexity": 1229.4852891468518}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243989819.92/warc/CC-MAIN-20210518094809-20210518124809-00188.warc.gz"}
https://www.clutchprep.com/physics/practice-problems/151495/1-the-power-required-to-exert-1-n-force-over-a-distance-of-1-m-in-1-second-isa-1
Power Video Lessons Concept # Problem: 1. The power required to exert 1 N force, over a distance of 1 m in 1 second isA) 1 W. B) 2 W. C) 1/3 W. D) 3 W. E) none of these.2. Do 100 J of work in 50 s and your power output isA) 1/2 W B) 2 W. C) 4 W. D) 50 W. E) 5,000 W. ###### FREE Expert Solution In this problem, we have to deal with force, distance, time, and power. We'll require an expression for the work and power. Power: $\overline{){\mathbf{P}}{\mathbf{=}}\frac{\mathbf{W}}{\mathbf{t}}}$ Work: $\overline{){\mathbf{W}}{\mathbf{=}}{\mathbf{F}}{\mathbf{d}}}$ 87% (30 ratings) ###### Problem Details 1. The power required to exert 1 N force, over a distance of 1 m in 1 second is A) 1 W. B) 2 W. C) 1/3 W. D) 3 W. E) none of these. 2. Do 100 J of work in 50 s and your power output is A) 1/2 W B) 2 W. C) 4 W. D) 50 W. E) 5,000 W. What scientific concept do you need to know in order to solve this problem? Our tutors have indicated that to solve this problem you will need to apply the Power concept. You can view video lessons to learn Power. Or if you need more Power practice, you can also practice Power practice problems.
2021-05-14 03:45:40
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 2, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7437231540679932, "perplexity": 2042.7788568051046}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243991737.39/warc/CC-MAIN-20210514025740-20210514055740-00129.warc.gz"}
https://www.physicsforums.com/threads/potential-of-a-charged-cone.813810/
# Potential of a charged cone 1. May 14, 2015 ### Quotes How to calculate the potential at the apex of uniformly charged right circular cone (charge only at the curved surface), having height "h" and radius "R" and lateral height "l" and change density sigma? 2. May 14, 2015 ### stevendaryl Staff Emeritus The potential at a point $\vec{r}_0$ can be obtained by integrating $\frac{dQ}{|\vec{r} - \vec{r}_0|}$ over the charge distribution. If you have a surface charge distribution, then $dQ = \sigma dA$ where $A$ is the area, and $\sigma$ is the charge density. So the challenge is to find a convenient coordinate system for computing $|\vec{r} - \vec{r}_0|$ and $dA$ and then just do the integral. Do you have any ideas about a convenient coordinate system for a cylinder? 3. May 14, 2015 ### Quotes Thanks for the hint... I tried it but its really getting quite difficult for me to choose a suitable area element for which i can calculate the potential and then integrating it over the whole area.... So can you please show me the steps to proceed... I need to get the question done today itself 4. May 14, 2015 ### stevendaryl Staff Emeritus Here's another hint: Suppose you split up the cone into narrow horizontal strips. Let $r$ be the distance of the strip from the apex. Let $dr$ be the width of the strip, and let $L$ be its length (the distance all the way around the strip). Then the area of the strip will be $dA = dr \cdot L$. So can you figure out $L$ in terms of $r$? 5. May 14, 2015 ### Quotes Are you asking me to take the element as small strips wrapped around the cone as a circle? 6. May 14, 2015 ### stevendaryl Staff Emeritus I'm getting closer and closer to just giving you the answer. But here's another clue: If you have a cone that measures $s$ down the side, and the side makes an angle $\theta$ relative to the vertical, then the area of the cone's surface is: $A = \pi s^2 sin(\theta)$ So you can take the derivative with respect to $s$ to get $dA$: $dA = 2 \pi s sin(\theta) ds$ That's the same as a little strip with width $ds$ and length $2 \pi s sin(\theta)$ So $dQ = \sigma\ dA = \sigma\ 2 \pi s sin(\theta) ds$ The distance from the strip to the apex is just $s$. So $\frac{dQ}{|\vec{r} - \vec{r}_0|} = \frac{dQ}{s}$ So do the integral of $\frac{dQ}{s}$ where $s$ goes from 0 to $l$. Can you figure out what $sin(\theta)$ must be? 7. May 14, 2015 ### Staff: Mentor @Quotes -- you need to start showing some effort, or this thread will be closed and deleted. We do not do students' work for them here at the PF. 8. May 14, 2015 ### Quotes Sorry for this.. It was my first day at the forum. But yes for sure i will keep that in mind. Thankyou
2018-02-24 06:55:47
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8641250729560852, "perplexity": 318.4392126257655}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-09/segments/1518891815435.68/warc/CC-MAIN-20180224053236-20180224073236-00487.warc.gz"}
https://www.physicsforums.com/threads/what-is-the-frequency-of-the-harmonic-potential.791494/
# What is the frequency of the harmonic potential? 1. Jan 10, 2015 ### KFC Hi there, I am reading an introduction on trapping atoms in space with magnetic potential. The article said the lab usually use a harmonic potential to trap the atoms and the potentials is in the form $\dfrac{m}{2}(\omega_x^2x^2 + \omega_y^2y^2 + \omega_z^2z^2)$ and $\omega_{x,y,z}$ has the unit of frequency. I wonder how do you understand the frequency in the potential from physical point of view. Why there is frequency? 2. Jan 10, 2015 ### ShayanJ Let's see what force that potential gives. For the x component, we have $F_x=-\frac{\partial V}{\partial x}=-m \omega_x^2 x$. But hey, that's Hooke's law!(With the spring constant $k_x=m \omega_x^2$.) So this potential is actually a non-isotropic harmonic oscillator potential and this is the reason you have frequencies in it. 3. Jan 10, 2015 ### KFC Thanks. So if can I say under that potential, it just like 3 harmonic oscillators along x, y and z each is oscillating at the frequency $\omega_x$, $\omega_y$ and $\omega_z$ independently? 4. Jan 10, 2015 ### ShayanJ No, its just one harmonic oscillator having three independent oscillations with different frequencies in different dimensions. But yes, in terms of degrees of freedom, its no different than having three independent harmonic oscillators with different frequencies. But you should note this resemblance may not be usable in the context you're considering. 5. Jan 10, 2015 ### KFC Got it. One more question. Usually, if you solve the harmonic oscillator with 3 oscillating frequencies along 3 different frequency, we will get a solution in 3-dimensional also. But if the frequency along x and y are way larger than the $\omega_z$. In some articles, I saw that people simply approximate the solution along the z direction only. I stuck on the explaining this approximation in physics. The first thing come to my mind is if the oscillator oscillating along x and y much faster than z, can we consider the system may see the average motion along x and y instead because of high frequency? So we could consider the amplitude of the solution along x and y just like a constant? Only the z direction depends on time? But before I find the explanation, I am also thing that if $\omega_z$ is way smaller than the $\omega_{x,t}$, can we consider the profile on the z direction is changing slowly in time, so we could consider the solution in z direction is a constant, the effective solution is along x and y direction. I know those two statements are contradictory. But I cannot tell which one (or all) is wrong. and why? 6. Jan 10, 2015 ### ShayanJ Can you point me to one of those "articles"?
2018-07-21 01:11:40
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 2, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8208045363426208, "perplexity": 506.8057588116554}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-30/segments/1531676592001.81/warc/CC-MAIN-20180720232914-20180721012914-00538.warc.gz"}
https://2022.help.altair.com/2022/hwsolvers/os/topics/solvers/os/force2_bulk_r.htm
# FORCE2 Bulk Data Entry Used to define a static force by specification of a value and four grid points that determine the direction. It can also be used to define the EXCITEID field (Amplitude "A") of dynamic loads in RLOAD1, RLOAD2, TLOAD1 and TLOAD2 Bulk Data Entries. Additionally, the FORCE2 entry can be defined as Follower Loads in Large Displacement Nonlinear Analysis. ## Format (1) (2) (3) (4) (5) (6) (7) (8) (9) (10) FORCE2 SID G F G1 G2 G3 G4 ## Example (1) (2) (3) (4) (5) (6) (7) (8) (9) (10) FORCE2 6 13 -2.93 16 13 18 19 ## Definitions Field Contents SI Unit Example (Integer > 0) G Loaded Grid point identification number. (Integer > 0) F Value of force. (Real) Gi Grid point identification numbers. (Integer > 0; G1 and G2 may not be coincident; G3 and G4 cannot be coincident.) $\stackrel{\to }{\text{f}}=\text{F}\stackrel{\to }{\text{N}}$ Where, $\stackrel{\to }{\text{N}}$ is a unit vector parallel to a vector calculated by the cross product of vectors from G1 to G2 and G3 to G4.
2023-02-01 03:38:30
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 2, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7440284490585327, "perplexity": 2609.4678279719597}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499899.9/warc/CC-MAIN-20230201013650-20230201043650-00005.warc.gz"}
https://mathematics.huji.ac.il/event/combinatorics-michael-simkin-huji?ref_tid=3830
# Combinatorics: Michael Simkin (HUJI) Title: A randomized construction of high girth regular graphs Abstract: We describe a new random greedy algorithm for generating regular graphs of high girth: Let $k > 2$ and $0 < c < 1$ be fixed. Let $n$ be even and set $g = c \log_{k-1} (n)$. Begin with a Hamilton cycle $G$ on $n$ vertices. As long as the smallest degree $\delta (G)<k$, choose, uniformly at random, two vertices $u,v \in V(G)$ of degree $\delta(G)$ whose distance is at least $g-1$. If there are no such vertex pairs, abort. Otherwise, add the edge $uv$ to $E(G)$. We show that with high probability this algorithm yields a $k$-regular graph with girth at least $g$. Our analysis also implies that there are $( \Omega (n) )^{kn/2}$ $k$-regular $n$-vertex graphs with girth at least $g$. This is joint work with Nati Linial. ## Date: Mon, 13/01/2020 - 10:00 to 12:00 ## Location: C-400, CS building
2021-01-18 11:34:37
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9001491069793701, "perplexity": 302.3948724626624}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610703514495.52/warc/CC-MAIN-20210118092350-20210118122350-00000.warc.gz"}
https://socratic.org/questions/given-that-p-x-x-4-ax-3-x-2-bx-12-has-factors-x-2-and-and-x-1-how-do-you-solve-t
# Given that P(x) = x^4 + ax^3 - x^2 + bx - 12 has factors x - 2 and and x + 1, how do you solve the equation P(x) = 0? This equation has $4$ solutions: ${x}_{1} = - 3$, ${x}_{2} = - 2$, ${x}_{3} = - 1$ and ${x}_{4} = 2$ According to Viete's Theorem if $P \left(x\right)$ has a factor of $\left(x - a\right)$ then $a$ is a root of this polynomial, so in this case this polynomial has 2 roots: ${x}_{3} = - 1$ and ${x}_{4} = 2$. To find the other roots you have to divide $P \left(x\right)$ by $\left(x - 2\right) \left(x + 1\right)$. The result will be: ${x}^{2} + 5 x + 6$.
2020-03-28 15:20:20
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 13, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8006737232208252, "perplexity": 93.98623428366346}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370491998.11/warc/CC-MAIN-20200328134227-20200328164227-00158.warc.gz"}