code
stringlengths
5
1.03M
repo_name
stringlengths
5
90
path
stringlengths
4
158
license
stringclasses
15 values
size
int64
5
1.03M
n_ast_errors
int64
0
53.9k
ast_max_depth
int64
2
4.17k
n_whitespaces
int64
0
365k
n_ast_nodes
int64
3
317k
n_ast_terminals
int64
1
171k
n_ast_nonterminals
int64
1
146k
loc
int64
-1
37.3k
cycloplexity
int64
-1
1.31k
{-# LANGUAGE Haskell2010, OverloadedStrings #-} {-# LINE 1 "Network/Wai/Middleware/AddHeaders.hs" #-} -- | -- -- Since 3.0.3 module Network.Wai.Middleware.AddHeaders ( addHeaders ) where import Network.HTTP.Types (Header) import Network.Wai (Middleware, modifyResponse, mapResponseHeaders) import Network.Wai.Internal (Response(..)) import Data.ByteString (ByteString) import qualified Data.CaseInsensitive as CI import Control.Arrow (first) addHeaders :: [(ByteString, ByteString)] -> Middleware -- ^ Prepend a list of headers without any checks -- -- Since 3.0.3 addHeaders h = modifyResponse $ addHeaders' (map (first CI.mk) h) addHeaders' :: [Header] -> Response -> Response addHeaders' h = mapResponseHeaders (\hs -> h ++ hs)
phischu/fragnix
tests/packages/scotty/Network.Wai.Middleware.AddHeaders.hs
bsd-3-clause
760
0
11
119
179
109
70
14
1
{-# LANGUAGE MultiParamTypeClasses, TypeSynonymInstances, FlexibleInstances #-} {- | Module : $Header$ Description : COL instance of class Logic Copyright : (c) Till Mossakowski, Uni Bremen 2002-2004 License : GPLv2 or higher, see LICENSE.txt Maintainer : [email protected] Stability : provisional Portability : non-portable (via Logic) COL instance of class Logic -} module COL.Logic_COL where import COL.AS_COL import COL.COLSign import COL.ATC_COL () import COL.Parse_AS () import COL.StatAna import COL.Print_AS () import CASL.Sign import CASL.StaticAna import CASL.MixfixParser import CASL.Morphism import CASL.SymbolMapAnalysis import CASL.AS_Basic_CASL import CASL.Parse_AS_Basic import CASL.MapSentence import CASL.SymbolParser import CASL.Logic_CASL () import Logic.Logic data COL = COL deriving Show instance Language COL where description _ = "COLCASL extends CASL by constructors and observers" type C_BASIC_SPEC = BASIC_SPEC () COL_SIG_ITEM () type CSign = Sign () COLSign type COLMor = Morphism () COLSign (DefMorExt COLSign) type COLFORMULA = FORMULA () instance SignExtension COLSign where isSubSignExtension = isSubCOLSign instance Syntax COL C_BASIC_SPEC Symbol SYMB_ITEMS SYMB_MAP_ITEMS where parse_basic_spec COL = Just $ basicSpec col_reserved_words parse_symb_items COL = Just $ symbItems col_reserved_words parse_symb_map_items COL = Just $ symbMapItems col_reserved_words instance Sentences COL COLFORMULA CSign COLMor Symbol where map_sen COL m = return . mapSen (const id) m sym_of COL = symOf symmap_of COL = morphismToSymbMap sym_name COL = symName instance StaticAnalysis COL C_BASIC_SPEC COLFORMULA SYMB_ITEMS SYMB_MAP_ITEMS CSign COLMor Symbol RawSymbol where basic_analysis COL = Just $ basicAnalysis (const return) (const return) ana_COL_SIG_ITEM emptyMix stat_symb_map_items COL = statSymbMapItems stat_symb_items COL = statSymbItems symbol_to_raw COL = symbolToRaw id_to_raw COL = idToRaw matches COL = CASL.Morphism.matches empty_signature COL = emptySign emptyCOLSign signature_union COL sigma1 = return . addSig addCOLSign sigma1 morphism_union COL = plainMorphismUnion addCOLSign final_union COL = finalUnion addCOLSign is_subsig COL = isSubSig isSubCOLSign subsig_inclusion COL = sigInclusion emptyMorExt cogenerated_sign COL = cogeneratedSign emptyMorExt generated_sign COL = generatedSign emptyMorExt induced_from_morphism COL = inducedFromMorphism emptyMorExt induced_from_to_morphism COL = inducedFromToMorphism emptyMorExt isSubCOLSign diffCOLSign instance Logic COL () C_BASIC_SPEC COLFORMULA SYMB_ITEMS SYMB_MAP_ITEMS CSign COLMor Symbol RawSymbol () where empty_proof_tree _ = ()
keithodulaigh/Hets
COL/Logic_COL.hs
gpl-2.0
3,110
0
9
791
596
307
289
70
0
{-# LANGUAGE DataKinds #-} {-# LANGUAGE DeriveGeneric #-} {-# LANGUAGE FlexibleInstances #-} {-# LANGUAGE GeneralizedNewtypeDeriving #-} {-# LANGUAGE LambdaCase #-} {-# LANGUAGE NoImplicitPrelude #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE RecordWildCards #-} {-# LANGUAGE TypeFamilies #-} {-# OPTIONS_GHC -fno-warn-unused-imports #-} -- Module : Network.AWS.Redshift.CreateTags -- Copyright : (c) 2013-2014 Brendan Hay <[email protected]> -- License : This Source Code Form is subject to the terms of -- the Mozilla Public License, v. 2.0. -- A copy of the MPL can be found in the LICENSE file or -- you can obtain it at http://mozilla.org/MPL/2.0/. -- Maintainer : Brendan Hay <[email protected]> -- Stability : experimental -- Portability : non-portable (GHC extensions) -- -- Derived from AWS service descriptions, licensed under Apache 2.0. -- | Adds one or more tags to a specified resource. -- -- A resource can have up to 10 tags. If you try to create more than 10 tags -- for a resource, you will receive an error and the attempt will fail. -- -- If you specify a key that already exists for the resource, the value for -- that key will be updated with the new value. -- -- <http://docs.aws.amazon.com/redshift/latest/APIReference/API_CreateTags.html> module Network.AWS.Redshift.CreateTags ( -- * Request CreateTags -- ** Request constructor , createTags -- ** Request lenses , ctResourceName , ctTags -- * Response , CreateTagsResponse -- ** Response constructor , createTagsResponse ) where import Network.AWS.Prelude import Network.AWS.Request.Query import Network.AWS.Redshift.Types import qualified GHC.Exts data CreateTags = CreateTags { _ctResourceName :: Text , _ctTags :: List "member" Tag } deriving (Eq, Read, Show) -- | 'CreateTags' constructor. -- -- The fields accessible through corresponding lenses are: -- -- * 'ctResourceName' @::@ 'Text' -- -- * 'ctTags' @::@ ['Tag'] -- createTags :: Text -- ^ 'ctResourceName' -> CreateTags createTags p1 = CreateTags { _ctResourceName = p1 , _ctTags = mempty } -- | The Amazon Resource Name (ARN) to which you want to add the tag or tags. For -- example, 'arn:aws:redshift:us-east-1:123456789:cluster:t1'. ctResourceName :: Lens' CreateTags Text ctResourceName = lens _ctResourceName (\s a -> s { _ctResourceName = a }) -- | One or more name/value pairs to add as tags to the specified resource. Each -- tag name is passed in with the parameter 'tag-key' and the corresponding value -- is passed in with the parameter 'tag-value'. The 'tag-key' and 'tag-value' -- parameters are separated by a colon (:). Separate multiple tags with a space. -- For example, '--tags "tag-key"="owner":"tag-value"="admin""tag-key"="environment":"tag-value"="test""tag-key"="version":"tag-value"="1.0"'. ctTags :: Lens' CreateTags [Tag] ctTags = lens _ctTags (\s a -> s { _ctTags = a }) . _List data CreateTagsResponse = CreateTagsResponse deriving (Eq, Ord, Read, Show, Generic) -- | 'CreateTagsResponse' constructor. createTagsResponse :: CreateTagsResponse createTagsResponse = CreateTagsResponse instance ToPath CreateTags where toPath = const "/" instance ToQuery CreateTags where toQuery CreateTags{..} = mconcat [ "ResourceName" =? _ctResourceName , "Tags" =? _ctTags ] instance ToHeaders CreateTags instance AWSRequest CreateTags where type Sv CreateTags = Redshift type Rs CreateTags = CreateTagsResponse request = post "CreateTags" response = nullResponse CreateTagsResponse
kim/amazonka
amazonka-redshift/gen/Network/AWS/Redshift/CreateTags.hs
mpl-2.0
3,796
0
10
837
405
251
154
51
1
{-# LANGUAGE OverloadedStrings #-} module Main where import Control.Applicative ((<$>)) import Control.Monad (forM_) import qualified Data.Aeson as Json import qualified Data.ByteString.Lazy as BS import qualified Data.Text.Lazy as Text import qualified Data.Text.Lazy.Encoding as Text import qualified Data.Text.Lazy.IO as Text import qualified System.Directory as Dir import qualified System.Environment as Env import System.Exit (exitFailure) import System.FilePath ((</>), takeExtension) import System.IO (hPutStrLn, stderr) import qualified Elm.Docs as Docs import Elm.Utils ((|>)) main :: IO () main = do allArgs <- Env.getArgs case allArgs of [pkg] -> do fixAllDocs pkg fixModuleDocs pkg _ -> do hPutStrLn stderr "Expecting a path to a particular version of a package" exitFailure -- FIX documentation.json fixAllDocs :: FilePath -> IO () fixAllDocs pkg = do maybeValue <- Json.decode <$> BS.readFile (pkg </> "documentation.json") writeAsJson (pkg </> "documentation.json") (maybeValue :: Maybe [Docs.Documentation]) -- FIX docs/*.json fixModuleDocs :: FilePath -> IO () fixModuleDocs pkg = do moduleDocs <- filter (\path -> takeExtension path == ".json") <$> Dir.getDirectoryContents (pkg </> "docs") forM_ moduleDocs $ \file -> do maybeValue <- Json.decode <$> BS.readFile (pkg </> "docs" </> file) writeAsJson (pkg </> "docs" </> file) (maybeValue :: Maybe Docs.Documentation) -- WRITE JSON writeAsJson :: (Json.ToJSON a) => FilePath -> Maybe a -> IO () writeAsJson file maybeValue = case maybeValue of Nothing -> do hPutStrLn stderr "Problem reading JSON" exitFailure Just value -> Docs.prettyJson value |> Text.decodeUtf8 |> Text.replace "\\u003e" ">" |> Text.writeFile file
Dedoig/package.elm-lang.org
upgrade-docs.hs
bsd-3-clause
1,977
0
16
519
535
291
244
53
2
{- (c) The University of Glasgow 2006 (c) The AQUA Project, Glasgow University, 1994-1998 Core-syntax unfoldings Unfoldings (which can travel across module boundaries) are in Core syntax (namely @CoreExpr@s). The type @Unfolding@ sits ``above'' simply-Core-expressions unfoldings, capturing ``higher-level'' things we know about a binding, usually things that the simplifier found out (e.g., ``it's a literal''). In the corner of a @CoreUnfolding@ unfolding, you will find, unsurprisingly, a Core expression. -} {-# LANGUAGE CPP #-} module ETA.Core.CoreUnfold ( Unfolding, UnfoldingGuidance, -- Abstract types noUnfolding, mkImplicitUnfolding, mkUnfolding, mkCoreUnfolding, mkTopUnfolding, mkSimpleUnfolding, mkWorkerUnfolding, mkInlineUnfolding, mkInlinableUnfolding, mkWwInlineRule, mkCompulsoryUnfolding, mkDFunUnfolding, specUnfolding, ArgSummary(..), couldBeSmallEnoughToInline, inlineBoringOk, certainlyWillInline, smallEnoughToInline, callSiteInline, CallCtxt(..), -- Reexport from CoreSubst (it only live there so it can be used -- by the Very Simple Optimiser) exprIsConApp_maybe, exprIsLiteral_maybe ) where #include "HsVersions.h" import ETA.Main.DynFlags import ETA.Core.CoreSyn import ETA.Core.PprCore () -- Instances import ETA.SimplCore.OccurAnal ( occurAnalyseExpr ) import ETA.Core.CoreSubst hiding( substTy ) import ETA.Core.CoreArity ( manifestArity, exprBotStrictness_maybe ) import ETA.Core.CoreUtils import ETA.BasicTypes.Id import ETA.BasicTypes.DataCon import ETA.BasicTypes.Literal import ETA.Prelude.PrimOp import ETA.BasicTypes.IdInfo import ETA.BasicTypes.BasicTypes ( Arity ) import ETA.Types.Type import ETA.Prelude.PrelNames import ETA.Prelude.TysPrim ( realWorldStatePrimTy ) import ETA.Utils.Bag import ETA.Utils.Util import ETA.Utils.FastTypes import ETA.Utils.FastString import ETA.Utils.Outputable import ETA.Prelude.ForeignCall import qualified Data.ByteString as BS import Data.Maybe {- ************************************************************************ * * \subsection{Making unfoldings} * * ************************************************************************ -} mkTopUnfolding :: DynFlags -> Bool -> CoreExpr -> Unfolding mkTopUnfolding dflags = mkUnfolding dflags InlineRhs True {- Top level -} mkImplicitUnfolding :: DynFlags -> CoreExpr -> Unfolding -- For implicit Ids, do a tiny bit of optimising first mkImplicitUnfolding dflags expr = mkTopUnfolding dflags False (simpleOptExpr expr) -- Note [Top-level flag on inline rules] -- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ -- Slight hack: note that mk_inline_rules conservatively sets the -- top-level flag to True. It gets set more accurately by the simplifier -- Simplify.simplUnfolding. mkSimpleUnfolding :: DynFlags -> CoreExpr -> Unfolding mkSimpleUnfolding dflags = mkUnfolding dflags InlineRhs False False mkDFunUnfolding :: [Var] -> DataCon -> [CoreExpr] -> Unfolding mkDFunUnfolding bndrs con ops = DFunUnfolding { df_bndrs = bndrs , df_con = con , df_args = map occurAnalyseExpr ops } -- See Note [Occurrrence analysis of unfoldings] mkWwInlineRule :: CoreExpr -> Arity -> Unfolding mkWwInlineRule expr arity = mkCoreUnfolding InlineStable True (simpleOptExpr expr) (UnfWhen { ug_arity = arity, ug_unsat_ok = unSaturatedOk , ug_boring_ok = boringCxtNotOk }) mkCompulsoryUnfolding :: CoreExpr -> Unfolding mkCompulsoryUnfolding expr -- Used for things that absolutely must be unfolded = mkCoreUnfolding InlineCompulsory True (simpleOptExpr expr) (UnfWhen { ug_arity = 0 -- Arity of unfolding doesn't matter , ug_unsat_ok = unSaturatedOk, ug_boring_ok = boringCxtOk }) mkWorkerUnfolding :: DynFlags -> (CoreExpr -> CoreExpr) -> Unfolding -> Unfolding -- See Note [Worker-wrapper for INLINABLE functions] in WorkWrap mkWorkerUnfolding dflags work_fn (CoreUnfolding { uf_src = src, uf_tmpl = tmpl , uf_is_top = top_lvl }) | isStableSource src = mkCoreUnfolding src top_lvl new_tmpl guidance where new_tmpl = simpleOptExpr (work_fn tmpl) guidance = calcUnfoldingGuidance dflags new_tmpl mkWorkerUnfolding _ _ _ = noUnfolding mkInlineUnfolding :: Maybe Arity -> CoreExpr -> Unfolding mkInlineUnfolding mb_arity expr = mkCoreUnfolding InlineStable True -- Note [Top-level flag on inline rules] expr' guide where expr' = simpleOptExpr expr guide = case mb_arity of Nothing -> UnfWhen { ug_arity = manifestArity expr' , ug_unsat_ok = unSaturatedOk , ug_boring_ok = boring_ok } Just arity -> UnfWhen { ug_arity = arity , ug_unsat_ok = needSaturated , ug_boring_ok = boring_ok } boring_ok = inlineBoringOk expr' mkInlinableUnfolding :: DynFlags -> CoreExpr -> Unfolding mkInlinableUnfolding dflags expr = mkUnfolding dflags InlineStable True is_bot expr' where expr' = simpleOptExpr expr is_bot = isJust (exprBotStrictness_maybe expr') specUnfolding :: DynFlags -> Subst -> [Var] -> [CoreExpr] -> Unfolding -> Unfolding -- See Note [Specialising unfoldings] specUnfolding _ subst new_bndrs spec_args df@(DFunUnfolding { df_bndrs = bndrs, df_con = con , df_args = args }) = ASSERT2( length bndrs >= length spec_args, ppr df $$ ppr spec_args $$ ppr new_bndrs ) mkDFunUnfolding (new_bndrs ++ extra_bndrs) con (map (substExpr spec_doc subst2) args) where subst1 = extendSubstList subst (bndrs `zip` spec_args) (subst2, extra_bndrs) = substBndrs subst1 (dropList spec_args bndrs) specUnfolding _dflags subst new_bndrs spec_args (CoreUnfolding { uf_src = src, uf_tmpl = tmpl , uf_is_top = top_lvl , uf_guidance = old_guidance }) | isStableSource src -- See Note [Specialising unfoldings] , UnfWhen { ug_arity = old_arity , ug_unsat_ok = unsat_ok , ug_boring_ok = boring_ok } <- old_guidance = let guidance = UnfWhen { ug_arity = old_arity - count isValArg spec_args + count isId new_bndrs , ug_unsat_ok = unsat_ok , ug_boring_ok = boring_ok } new_tmpl = simpleOptExpr $ mkLams new_bndrs $ mkApps (substExpr spec_doc subst tmpl) spec_args -- The beta-redexes created here will be simplified -- away by simplOptExpr in mkUnfolding in mkCoreUnfolding src top_lvl new_tmpl guidance specUnfolding _ _ _ _ _ = noUnfolding spec_doc :: SDoc spec_doc = ptext (sLit "specUnfolding") {- Note [Specialising unfoldings] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ When we specialise a function for some given type-class arguments, we use specUnfolding to specialise its unfolding. Some important points: * If the original function has a DFunUnfolding, the specialised one must do so too! Otherwise we lose the magic rules that make it interact with ClassOps * There is a bit of hack for INLINABLE functions: f :: Ord a => .... f = <big-rhs> {- INLINEABLE f #-} Now if we specialise f, should the specialised version still have an INLINEABLE pragma? If it does, we'll capture a specialised copy of <big-rhs> as its unfolding, and that probaby won't inline. But if we don't, the specialised version of <big-rhs> might be small enough to inline at a call site. This happens with Control.Monad.liftM3, and can cause a lot more allocation as a result (nofib n-body shows this). Moreover, keeping the INLINEABLE thing isn't much help, because the specialised function (probaby) isn't overloaded any more. Conclusion: drop the INLINEALE pragma. In practice what this means is: if a stable unfolding has UnfoldingGuidance of UnfWhen, we keep it (so the specialised thing too will always inline) if a stable unfolding has UnfoldingGuidance of UnfIfGoodArgs (which arises from INLINEABLE), we discard it -} mkCoreUnfolding :: UnfoldingSource -> Bool -> CoreExpr -> UnfoldingGuidance -> Unfolding -- Occurrence-analyses the expression before capturing it mkCoreUnfolding src top_lvl expr guidance = CoreUnfolding { uf_tmpl = occurAnalyseExpr expr, -- See Note [Occurrrence analysis of unfoldings] uf_src = src, uf_is_top = top_lvl, uf_is_value = exprIsHNF expr, uf_is_conlike = exprIsConLike expr, uf_is_work_free = exprIsWorkFree expr, uf_expandable = exprIsExpandable expr, uf_guidance = guidance } mkUnfolding :: DynFlags -> UnfoldingSource -> Bool -> Bool -> CoreExpr -> Unfolding -- Calculates unfolding guidance -- Occurrence-analyses the expression before capturing it mkUnfolding dflags src top_lvl is_bottoming expr | top_lvl && is_bottoming , not (exprIsTrivial expr) = NoUnfolding -- See Note [Do not inline top-level bottoming functions] | otherwise = CoreUnfolding { uf_tmpl = occurAnalyseExpr expr, -- See Note [Occurrrence analysis of unfoldings] uf_src = src, uf_is_top = top_lvl, uf_is_value = exprIsHNF expr, uf_is_conlike = exprIsConLike expr, uf_expandable = exprIsExpandable expr, uf_is_work_free = exprIsWorkFree expr, uf_guidance = guidance } where guidance = calcUnfoldingGuidance dflags expr -- NB: *not* (calcUnfoldingGuidance (occurAnalyseExpr expr))! -- See Note [Calculate unfolding guidance on the non-occ-anal'd expression] {- Note [Occurrence analysis of unfoldings] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ We do occurrence-analysis of unfoldings once and for all, when the unfolding is built, rather than each time we inline them. But given this decision it's vital that we do *always* do it. Consider this unfolding \x -> letrec { f = ...g...; g* = f } in body where g* is (for some strange reason) the loop breaker. If we don't occ-anal it when reading it in, we won't mark g as a loop breaker, and we may inline g entirely in body, dropping its binding, and leaving the occurrence in f out of scope. This happened in Trac #8892, where the unfolding in question was a DFun unfolding. But more generally, the simplifier is designed on the basis that it is looking at occurrence-analysed expressions, so better ensure that they acutally are. Note [Calculate unfolding guidance on the non-occ-anal'd expression] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Notice that we give the non-occur-analysed expression to calcUnfoldingGuidance. In some ways it'd be better to occur-analyse first; for example, sometimes during simplification, there's a large let-bound thing which has been substituted, and so is now dead; so 'expr' contains two copies of the thing while the occurrence-analysed expression doesn't. Nevertheless, we *don't* and *must not* occ-analyse before computing the size because a) The size computation bales out after a while, whereas occurrence analysis does not. b) Residency increases sharply if you occ-anal first. I'm not 100% sure why, but it's a large effect. Compiling Cabal went from residency of 534M to over 800M with this one change. This can occasionally mean that the guidance is very pessimistic; it gets fixed up next round. And it should be rare, because large let-bound things that are dead are usually caught by preInlineUnconditionally ************************************************************************ * * \subsection{The UnfoldingGuidance type} * * ************************************************************************ -} inlineBoringOk :: CoreExpr -> Bool -- See Note [INLINE for small functions] -- True => the result of inlining the expression is -- no bigger than the expression itself -- eg (\x y -> f y x) -- This is a quick and dirty version. It doesn't attempt -- to deal with (\x y z -> x (y z)) -- The really important one is (x `cast` c) inlineBoringOk e = go 0 e where go :: Int -> CoreExpr -> Bool go credit (Lam x e) | isId x = go (credit+1) e | otherwise = go credit e go credit (App f (Type {})) = go credit f go credit (App f a) | credit > 0 , exprIsTrivial a = go (credit-1) f go credit (Tick _ e) = go credit e -- dubious go credit (Cast e _) = go credit e go _ (Var {}) = boringCxtOk go _ _ = boringCxtNotOk calcUnfoldingGuidance :: DynFlags -> CoreExpr -- Expression to look at -> UnfoldingGuidance calcUnfoldingGuidance dflags (Tick t expr) | not (tickishIsCode t) -- non-code ticks don't matter for unfolding = calcUnfoldingGuidance dflags expr calcUnfoldingGuidance dflags expr = case sizeExpr dflags (iUnbox bOMB_OUT_SIZE) val_bndrs body of TooBig -> UnfNever SizeIs size cased_bndrs scrut_discount | uncondInline expr n_val_bndrs (iBox size) -> UnfWhen { ug_unsat_ok = unSaturatedOk , ug_boring_ok = boringCxtOk , ug_arity = n_val_bndrs } -- Note [INLINE for small functions] | otherwise -> UnfIfGoodArgs { ug_args = map (mk_discount cased_bndrs) val_bndrs , ug_size = iBox size , ug_res = iBox scrut_discount } where (bndrs, body) = collectBinders expr bOMB_OUT_SIZE = ufCreationThreshold dflags -- Bomb out if size gets bigger than this val_bndrs = filter isId bndrs n_val_bndrs = length val_bndrs mk_discount :: Bag (Id,Int) -> Id -> Int mk_discount cbs bndr = foldlBag combine 0 cbs where combine acc (bndr', disc) | bndr == bndr' = acc `plus_disc` disc | otherwise = acc plus_disc :: Int -> Int -> Int plus_disc | isFunTy (idType bndr) = max | otherwise = (+) -- See Note [Function and non-function discounts] {- Note [Computing the size of an expression] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ The basic idea of sizeExpr is obvious enough: count nodes. But getting the heuristics right has taken a long time. Here's the basic strategy: * Variables, literals: 0 (Exception for string literals, see litSize.) * Function applications (f e1 .. en): 1 + #value args * Constructor applications: 1, regardless of #args * Let(rec): 1 + size of components * Note, cast: 0 Examples Size Term -------------- 0 42# 0 x 0 True 2 f x 1 Just x 4 f (g x) Notice that 'x' counts 0, while (f x) counts 2. That's deliberate: there's a function call to account for. Notice also that constructor applications are very cheap, because exposing them to a caller is so valuable. [25/5/11] All sizes are now multiplied by 10, except for primops (which have sizes like 1 or 4. This makes primops look fantastically cheap, and seems to be almost unversally beneficial. Done partly as a result of #4978. Note [Do not inline top-level bottoming functions] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ The FloatOut pass has gone to some trouble to float out calls to 'error' and similar friends. See Note [Bottoming floats] in SetLevels. Do not re-inline them! But we *do* still inline if they are very small (the uncondInline stuff). Note [INLINE for small functions] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Consider {-# INLINE f #-} f x = Just x g y = f y Then f's RHS is no larger than its LHS, so we should inline it into even the most boring context. In general, f the function is sufficiently small that its body is as small as the call itself, the inline unconditionally, regardless of how boring the context is. Things to note: (1) We inline *unconditionally* if inlined thing is smaller (using sizeExpr) than the thing it's replacing. Notice that (f x) --> (g 3) -- YES, unconditionally (f x) --> x : [] -- YES, *even though* there are two -- arguments to the cons x --> g 3 -- NO x --> Just v -- NO It's very important not to unconditionally replace a variable by a non-atomic term. (2) We do this even if the thing isn't saturated, else we end up with the silly situation that f x y = x ...map (f 3)... doesn't inline. Even in a boring context, inlining without being saturated will give a lambda instead of a PAP, and will be more efficient at runtime. (3) However, when the function's arity > 0, we do insist that it has at least one value argument at the call site. (This check is made in the UnfWhen case of callSiteInline.) Otherwise we find this: f = /\a \x:a. x d = /\b. MkD (f b) If we inline f here we get d = /\b. MkD (\x:b. x) and then prepareRhs floats out the argument, abstracting the type variables, so we end up with the original again! (4) We must be much more cautious about arity-zero things. Consider let x = y +# z in ... In *size* terms primops look very small, because the generate a single instruction, but we do not want to unconditionally replace every occurrence of x with (y +# z). So we only do the unconditional-inline thing for *trivial* expressions. NB: you might think that PostInlineUnconditionally would do this but it doesn't fire for top-level things; see SimplUtils Note [Top level and postInlineUnconditionally] -} uncondInline :: CoreExpr -> Arity -> Int -> Bool -- Inline unconditionally if there no size increase -- Size of call is arity (+1 for the function) -- See Note [INLINE for small functions] uncondInline rhs arity size | arity > 0 = size <= 10 * (arity + 1) -- See Note [INLINE for small functions] (1) | otherwise = exprIsTrivial rhs -- See Note [INLINE for small functions] (4) sizeExpr :: DynFlags -> FastInt -- Bomb out if it gets bigger than this -> [Id] -- Arguments; we're interested in which of these -- get case'd -> CoreExpr -> ExprSize -- Note [Computing the size of an expression] sizeExpr dflags bOMB_OUT_SIZE top_args expr = size_up expr where size_up (Cast e _) = size_up e size_up (Tick _ e) = size_up e size_up (Type _) = sizeZero -- Types cost nothing size_up (Coercion _) = sizeZero size_up (Lit lit) = sizeN (litSize lit) size_up (Var f) | isRealWorldId f = sizeZero -- Make sure we get constructor discounts even -- on nullary constructors | otherwise = size_up_call f [] 0 size_up (App fun arg) | isTyCoArg arg = size_up fun | otherwise = size_up arg `addSizeNSD` size_up_app fun [arg] (if isRealWorldExpr arg then 1 else 0) size_up (Lam b e) | isId b && not (isRealWorldId b) = lamScrutDiscount dflags (size_up e `addSizeN` 10) | otherwise = size_up e size_up (Let (NonRec binder rhs) body) = size_up rhs `addSizeNSD` size_up body `addSizeN` (if isUnLiftedType (idType binder) then 0 else 10) -- For the allocation -- If the binder has an unlifted type there is no allocation size_up (Let (Rec pairs) body) = foldr (addSizeNSD . size_up . snd) (size_up body `addSizeN` (10 * length pairs)) -- (length pairs) for the allocation pairs size_up (Case (Var v) _ _ alts) | v `elem` top_args -- We are scrutinising an argument variable = alts_size (foldr addAltSize sizeZero alt_sizes) (foldr maxSize sizeZero alt_sizes) -- Good to inline if an arg is scrutinised, because -- that may eliminate allocation in the caller -- And it eliminates the case itself where alt_sizes = map size_up_alt alts -- alts_size tries to compute a good discount for -- the case when we are scrutinising an argument variable alts_size (SizeIs tot tot_disc tot_scrut) -- Size of all alternatives (SizeIs max _ _) -- Size of biggest alternative = SizeIs tot (unitBag (v, iBox (_ILIT(20) +# tot -# max)) `unionBags` tot_disc) tot_scrut -- If the variable is known, we produce a discount that -- will take us back to 'max', the size of the largest alternative -- The 1+ is a little discount for reduced allocation in the caller -- -- Notice though, that we return tot_disc, the total discount from -- all branches. I think that's right. alts_size tot_size _ = tot_size size_up (Case e _ _ alts) = size_up e `addSizeNSD` foldr (addAltSize . size_up_alt) case_size alts where case_size | is_inline_scrut e, not (lengthExceeds alts 1) = sizeN (-10) | otherwise = sizeZero -- Normally we don't charge for the case itself, but -- we charge one per alternative (see size_up_alt, -- below) to account for the cost of the info table -- and comparisons. -- -- However, in certain cases (see is_inline_scrut -- below), no code is generated for the case unless -- there are multiple alts. In these cases we -- subtract one, making the first alt free. -- e.g. case x# +# y# of _ -> ... should cost 1 -- case touch# x# of _ -> ... should cost 0 -- (see #4978) -- -- I would like to not have the "not (lengthExceeds alts 1)" -- condition above, but without that some programs got worse -- (spectral/hartel/event and spectral/para). I don't fully -- understand why. (SDM 24/5/11) -- unboxed variables, inline primops and unsafe foreign calls -- are all "inline" things: is_inline_scrut (Var v) = isUnLiftedType (idType v) is_inline_scrut scrut | (Var f, _) <- collectArgs scrut = case idDetails f of FCallId fc -> not (isSafeForeignCall fc) PrimOpId op -> not (primOpOutOfLine op) _other -> False | otherwise = False ------------ -- size_up_app is used when there's ONE OR MORE value args size_up_app (App fun arg) args voids | isTyCoArg arg = size_up_app fun args voids | isRealWorldExpr arg = size_up_app fun (arg:args) (voids + 1) | otherwise = size_up arg `addSizeNSD` size_up_app fun (arg:args) voids size_up_app (Var fun) args voids = size_up_call fun args voids size_up_app (Tick _ expr) args voids = size_up_app expr args voids size_up_app other args voids = size_up other `addSizeN` (length args - voids) ------------ size_up_call :: Id -> [CoreExpr] -> Int -> ExprSize size_up_call fun val_args voids = case idDetails fun of FCallId _ -> sizeN (10 * (1 + length val_args)) DataConWorkId dc -> conSize dc (length val_args) PrimOpId op -> primOpSize op (length val_args) ClassOpId _ -> classOpSize dflags top_args val_args _ -> funSize dflags top_args fun (length val_args) voids ------------ size_up_alt (_con, _bndrs, rhs) = size_up rhs `addSizeN` 10 -- Don't charge for args, so that wrappers look cheap -- (See comments about wrappers with Case) -- -- IMPORATANT: *do* charge 1 for the alternative, else we -- find that giant case nests are treated as practically free -- A good example is Foreign.C.Error.errrnoToIOError ------------ -- These addSize things have to be here because -- I don't want to give them bOMB_OUT_SIZE as an argument addSizeN TooBig _ = TooBig addSizeN (SizeIs n xs d) m = mkSizeIs bOMB_OUT_SIZE (n +# iUnbox m) xs d -- addAltSize is used to add the sizes of case alternatives addAltSize TooBig _ = TooBig addAltSize _ TooBig = TooBig addAltSize (SizeIs n1 xs d1) (SizeIs n2 ys d2) = mkSizeIs bOMB_OUT_SIZE (n1 +# n2) (xs `unionBags` ys) (d1 +# d2) -- Note [addAltSize result discounts] -- This variant ignores the result discount from its LEFT argument -- It's used when the second argument isn't part of the result addSizeNSD TooBig _ = TooBig addSizeNSD _ TooBig = TooBig addSizeNSD (SizeIs n1 xs _) (SizeIs n2 ys d2) = mkSizeIs bOMB_OUT_SIZE (n1 +# n2) (xs `unionBags` ys) d2 -- Ignore d1 isRealWorldId id = idType id `eqType` realWorldStatePrimTy -- an expression of type State# RealWorld must be a variable isRealWorldExpr (Var id) = isRealWorldId id isRealWorldExpr (Tick _ e) = isRealWorldExpr e isRealWorldExpr _ = False -- | Finds a nominal size of a string literal. litSize :: Literal -> Int -- Used by CoreUnfold.sizeExpr litSize (LitInteger {}) = 100 -- Note [Size of literal integers] litSize (MachStr str) = 10 + 10 * ((BS.length str + 3) `div` 4) -- If size could be 0 then @f "x"@ might be too small -- [Sept03: make literal strings a bit bigger to avoid fruitless -- duplication of little strings] litSize _other = 0 -- Must match size of nullary constructors -- Key point: if x |-> 4, then x must inline unconditionally -- (eg via case binding) classOpSize :: DynFlags -> [Id] -> [CoreExpr] -> ExprSize -- See Note [Conlike is interesting] classOpSize _ _ [] = sizeZero classOpSize dflags top_args (arg1 : other_args) = SizeIs (iUnbox size) arg_discount (_ILIT(0)) where size = 20 + (10 * length other_args) -- If the class op is scrutinising a lambda bound dictionary then -- give it a discount, to encourage the inlining of this function -- The actual discount is rather arbitrarily chosen arg_discount = case arg1 of Var dict | dict `elem` top_args -> unitBag (dict, ufDictDiscount dflags) _other -> emptyBag funSize :: DynFlags -> [Id] -> Id -> Int -> Int -> ExprSize -- Size for functions that are not constructors or primops -- Note [Function applications] funSize dflags top_args fun n_val_args voids | fun `hasKey` buildIdKey = buildSize | fun `hasKey` augmentIdKey = augmentSize | otherwise = SizeIs (iUnbox size) arg_discount (iUnbox res_discount) where some_val_args = n_val_args > 0 size | some_val_args = 10 * (1 + n_val_args - voids) | otherwise = 0 -- The 1+ is for the function itself -- Add 1 for each non-trivial arg; -- the allocation cost, as in let(rec) -- DISCOUNTS -- See Note [Function and non-function discounts] arg_discount | some_val_args && fun `elem` top_args = unitBag (fun, ufFunAppDiscount dflags) | otherwise = emptyBag -- If the function is an argument and is applied -- to some values, give it an arg-discount res_discount | idArity fun > n_val_args = ufFunAppDiscount dflags | otherwise = 0 -- If the function is partially applied, show a result discount conSize :: DataCon -> Int -> ExprSize conSize dc n_val_args | n_val_args == 0 = SizeIs (_ILIT(0)) emptyBag (_ILIT(10)) -- Like variables -- See Note [Unboxed tuple size and result discount] | isUnboxedTupleCon dc = SizeIs (_ILIT(0)) emptyBag (iUnbox (10 * (1 + n_val_args))) -- See Note [Constructor size and result discount] | otherwise = SizeIs (_ILIT(10)) emptyBag (iUnbox (10 * (1 + n_val_args))) {- Note [Constructor size and result discount] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Treat a constructors application as size 10, regardless of how many arguments it has; we are keen to expose them (and we charge separately for their args). We can't treat them as size zero, else we find that (Just x) has size 0, which is the same as a lone variable; and hence 'v' will always be replaced by (Just x), where v is bound to Just x. The "result discount" is applied if the result of the call is scrutinised (say by a case). For a constructor application that will mean the constructor application will disappear, so we don't need to charge it to the function. So the discount should at least match the cost of the constructor application, namely 10. But to give a bit of extra incentive we give a discount of 10*(1 + n_val_args). Simon M tried a MUCH bigger discount: (10 * (10 + n_val_args)), and said it was an "unambiguous win", but its terribly dangerous because a fuction with many many case branches, each finishing with a constructor, can have an arbitrarily large discount. This led to terrible code bloat: see Trac #6099. Note [Unboxed tuple size and result discount] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ However, unboxed tuples count as size zero. I found occasions where we had f x y z = case op# x y z of { s -> (# s, () #) } and f wasn't getting inlined. I tried giving unboxed tuples a *result discount* of zero (see the commented-out line). Why? When returned as a result they do not allocate, so maybe we don't want to charge so much for them If you have a non-zero discount here, we find that workers often get inlined back into wrappers, because it look like f x = case $wf x of (# a,b #) -> (a,b) and we are keener because of the case. However while this change shrank binary sizes by 0.5% it also made spectral/boyer allocate 5% more. All other changes were very small. So it's not a big deal but I didn't adopt the idea. Note [Function and non-function discounts] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ We want a discount if the function is applied. A good example is monadic combinators with continuation arguments, where inlining is quite important. But we don't want a big discount when a function is called many times (see the detailed comments with Trac #6048) because if the function is big it won't be inlined at its many call sites and no benefit results. Indeed, we can get exponentially big inlinings this way; that is what Trac #6048 is about. On the other hand, for data-valued arguments, if there are lots of case expressions in the body, each one will get smaller if we apply the function to a constructor application, so we *want* a big discount if the argument is scrutinised by many case expressions. Conclusion: - For functions, take the max of the discounts - For data values, take the sum of the discounts Note [Literal integer size] ~~~~~~~~~~~~~~~~~~~~~~~~~~~ Literal integers *can* be big (mkInteger [...coefficients...]), but need not be (S# n). We just use an aribitrary big-ish constant here so that, in particular, we don't inline top-level defns like n = S# 5 There's no point in doing so -- any optimisations will see the S# through n's unfolding. Nor will a big size inhibit unfoldings functions that mention a literal Integer, because the float-out pass will float all those constants to top level. -} primOpSize :: PrimOp -> Int -> ExprSize primOpSize op n_val_args = if primOpOutOfLine op then sizeN (op_size + n_val_args) else sizeN op_size where op_size = primOpCodeSize op buildSize :: ExprSize buildSize = SizeIs (_ILIT(0)) emptyBag (_ILIT(40)) -- We really want to inline applications of build -- build t (\cn -> e) should cost only the cost of e (because build will be inlined later) -- Indeed, we should add a result_discount becuause build is -- very like a constructor. We don't bother to check that the -- build is saturated (it usually is). The "-2" discounts for the \c n, -- The "4" is rather arbitrary. augmentSize :: ExprSize augmentSize = SizeIs (_ILIT(0)) emptyBag (_ILIT(40)) -- Ditto (augment t (\cn -> e) ys) should cost only the cost of -- e plus ys. The -2 accounts for the \cn -- When we return a lambda, give a discount if it's used (applied) lamScrutDiscount :: DynFlags -> ExprSize -> ExprSize lamScrutDiscount dflags (SizeIs n vs _) = SizeIs n vs (iUnbox (ufFunAppDiscount dflags)) lamScrutDiscount _ TooBig = TooBig {- Note [addAltSize result discounts] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ When adding the size of alternatives, we *add* the result discounts too, rather than take the *maximum*. For a multi-branch case, this gives a discount for each branch that returns a constructor, making us keener to inline. I did try using 'max' instead, but it makes nofib 'rewrite' and 'puzzle' allocate significantly more, and didn't make binary sizes shrink significantly either. Note [Discounts and thresholds] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Constants for discounts and thesholds are defined in main/DynFlags, all of form ufXxxx. They are: ufCreationThreshold At a definition site, if the unfolding is bigger than this, we may discard it altogether ufUseThreshold At a call site, if the unfolding, less discounts, is smaller than this, then it's small enough inline ufKeenessFactor Factor by which the discounts are multiplied before subtracting from size ufDictDiscount The discount for each occurrence of a dictionary argument as an argument of a class method. Should be pretty small else big functions may get inlined ufFunAppDiscount Discount for a function argument that is applied. Quite large, because if we inline we avoid the higher-order call. ufDearOp The size of a foreign call or not-dupable PrimOp Note [Function applications] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In a function application (f a b) - If 'f' is an argument to the function being analysed, and there's at least one value arg, record a FunAppDiscount for f - If the application if a PAP (arity > 2 in this example) record a *result* discount (because inlining with "extra" args in the call may mean that we now get a saturated application) Code for manipulating sizes -} data ExprSize = TooBig | SizeIs FastInt -- Size found !(Bag (Id,Int)) -- Arguments cased herein, and discount for each such FastInt -- Size to subtract if result is scrutinised -- by a case expression instance Outputable ExprSize where ppr TooBig = ptext (sLit "TooBig") ppr (SizeIs a _ c) = brackets (int (iBox a) <+> int (iBox c)) -- subtract the discount before deciding whether to bale out. eg. we -- want to inline a large constructor application into a selector: -- tup = (a_1, ..., a_99) -- x = case tup of ... -- mkSizeIs :: FastInt -> FastInt -> Bag (Id, Int) -> FastInt -> ExprSize mkSizeIs max n xs d | (n -# d) ># max = TooBig | otherwise = SizeIs n xs d maxSize :: ExprSize -> ExprSize -> ExprSize maxSize TooBig _ = TooBig maxSize _ TooBig = TooBig maxSize s1@(SizeIs n1 _ _) s2@(SizeIs n2 _ _) | n1 ># n2 = s1 | otherwise = s2 sizeZero :: ExprSize sizeN :: Int -> ExprSize sizeZero = SizeIs (_ILIT(0)) emptyBag (_ILIT(0)) sizeN n = SizeIs (iUnbox n) emptyBag (_ILIT(0)) {- ************************************************************************ * * \subsection[considerUnfolding]{Given all the info, do (not) do the unfolding} * * ************************************************************************ We use 'couldBeSmallEnoughToInline' to avoid exporting inlinings that we ``couldn't possibly use'' on the other side. Can be overridden w/ flaggery. Just the same as smallEnoughToInline, except that it has no actual arguments. -} couldBeSmallEnoughToInline :: DynFlags -> Int -> CoreExpr -> Bool couldBeSmallEnoughToInline dflags threshold rhs = case sizeExpr dflags (iUnbox threshold) [] body of TooBig -> False _ -> True where (_, body) = collectBinders rhs ---------------- smallEnoughToInline :: DynFlags -> Unfolding -> Bool smallEnoughToInline dflags (CoreUnfolding {uf_guidance = UnfIfGoodArgs {ug_size = size}}) = size <= ufUseThreshold dflags smallEnoughToInline _ _ = False ---------------- certainlyWillInline :: DynFlags -> Unfolding -> Maybe Unfolding -- Sees if the unfolding is pretty certain to inline -- If so, return a *stable* unfolding for it, that will always inline certainlyWillInline dflags unf@(CoreUnfolding { uf_guidance = guidance, uf_tmpl = expr }) = case guidance of UnfNever -> Nothing UnfWhen {} -> Just (unf { uf_src = InlineStable }) -- The UnfIfGoodArgs case seems important. If we w/w small functions -- binary sizes go up by 10%! (This is with SplitObjs.) I'm not totally -- sure whyy. UnfIfGoodArgs { ug_size = size, ug_args = args } | not (null args) -- See Note [certainlyWillInline: be careful of thunks] , let arity = length args , size - (10 * (arity + 1)) <= ufUseThreshold dflags -> Just (unf { uf_src = InlineStable , uf_guidance = UnfWhen { ug_arity = arity , ug_unsat_ok = unSaturatedOk , ug_boring_ok = inlineBoringOk expr } }) -- Note the "unsaturatedOk". A function like f = \ab. a -- will certainly inline, even if partially applied (f e), so we'd -- better make sure that the transformed inlining has the same property _ -> Nothing certainlyWillInline _ unf@(DFunUnfolding {}) = Just unf certainlyWillInline _ _ = Nothing {- Note [certainlyWillInline: be careful of thunks] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Don't claim that thunks will certainly inline, because that risks work duplication. Even if the work duplication is not great (eg is_cheap holds), it can make a big difference in an inner loop In Trac #5623 we found that the WorkWrap phase thought that y = case x of F# v -> F# (v +# v) was certainlyWillInline, so the addition got duplicated. ************************************************************************ * * \subsection{callSiteInline} * * ************************************************************************ This is the key function. It decides whether to inline a variable at a call site callSiteInline is used at call sites, so it is a bit more generous. It's a very important function that embodies lots of heuristics. A non-WHNF can be inlined if it doesn't occur inside a lambda, and occurs exactly once or occurs once in each branch of a case and is small If the thing is in WHNF, there's no danger of duplicating work, so we can inline if it occurs once, or is small NOTE: we don't want to inline top-level functions that always diverge. It just makes the code bigger. Tt turns out that the convenient way to prevent them inlining is to give them a NOINLINE pragma, which we do in StrictAnal.addStrictnessInfoToTopId -} callSiteInline :: DynFlags -> Id -- The Id -> Bool -- True <=> unfolding is active -> Bool -- True if there are are no arguments at all (incl type args) -> [ArgSummary] -- One for each value arg; True if it is interesting -> CallCtxt -- True <=> continuation is interesting -> Maybe CoreExpr -- Unfolding, if any data ArgSummary = TrivArg -- Nothing interesting | NonTrivArg -- Arg has structure | ValueArg -- Arg is a con-app or PAP -- ..or con-like. Note [Conlike is interesting] instance Outputable ArgSummary where ppr TrivArg = ptext (sLit "TrivArg") ppr NonTrivArg = ptext (sLit "NonTrivArg") ppr ValueArg = ptext (sLit "ValueArg") nonTriv :: ArgSummary -> Bool nonTriv TrivArg = False nonTriv _ = True data CallCtxt = BoringCtxt | RhsCtxt -- Rhs of a let-binding; see Note [RHS of lets] | DiscArgCtxt -- Argument of a fuction with non-zero arg discount | RuleArgCtxt -- We are somewhere in the argument of a function with rules | ValAppCtxt -- We're applied to at least one value arg -- This arises when we have ((f x |> co) y) -- Then the (f x) has argument 'x' but in a ValAppCtxt | CaseCtxt -- We're the scrutinee of a case -- that decomposes its scrutinee instance Outputable CallCtxt where ppr CaseCtxt = ptext (sLit "CaseCtxt") ppr ValAppCtxt = ptext (sLit "ValAppCtxt") ppr BoringCtxt = ptext (sLit "BoringCtxt") ppr RhsCtxt = ptext (sLit "RhsCtxt") ppr DiscArgCtxt = ptext (sLit "DiscArgCtxt") ppr RuleArgCtxt = ptext (sLit "RuleArgCtxt") callSiteInline dflags id active_unfolding lone_variable arg_infos cont_info = case idUnfolding id of -- idUnfolding checks for loop-breakers, returning NoUnfolding -- Things with an INLINE pragma may have an unfolding *and* -- be a loop breaker (maybe the knot is not yet untied) CoreUnfolding { uf_tmpl = unf_template, uf_is_top = is_top , uf_is_work_free = is_wf , uf_guidance = guidance, uf_expandable = is_exp } | active_unfolding -> tryUnfolding dflags id lone_variable arg_infos cont_info unf_template is_top is_wf is_exp guidance | otherwise -> traceInline dflags "Inactive unfolding:" (ppr id) Nothing NoUnfolding -> Nothing OtherCon {} -> Nothing DFunUnfolding {} -> Nothing -- Never unfold a DFun traceInline :: DynFlags -> String -> SDoc -> a -> a traceInline dflags str doc result | dopt Opt_D_dump_inlinings dflags && dopt Opt_D_verbose_core2core dflags = pprTrace str doc result | otherwise = result tryUnfolding :: DynFlags -> Id -> Bool -> [ArgSummary] -> CallCtxt -> CoreExpr -> Bool -> Bool -> Bool -> UnfoldingGuidance -> Maybe CoreExpr tryUnfolding dflags id lone_variable arg_infos cont_info unf_template is_top is_wf is_exp guidance = case guidance of UnfNever -> traceInline dflags str (ptext (sLit "UnfNever")) Nothing UnfWhen { ug_arity = uf_arity, ug_unsat_ok = unsat_ok, ug_boring_ok = boring_ok } | enough_args && (boring_ok || some_benefit) -- See Note [INLINE for small functions (3)] -> traceInline dflags str (mk_doc some_benefit empty True) (Just unf_template) | otherwise -> traceInline dflags str (mk_doc some_benefit empty False) Nothing where some_benefit = calc_some_benefit uf_arity enough_args = (n_val_args >= uf_arity) || (unsat_ok && n_val_args > 0) UnfIfGoodArgs { ug_args = arg_discounts, ug_res = res_discount, ug_size = size } | is_wf && some_benefit && small_enough -> traceInline dflags str (mk_doc some_benefit extra_doc True) (Just unf_template) | otherwise -> traceInline dflags str (mk_doc some_benefit extra_doc False) Nothing where some_benefit = calc_some_benefit (length arg_discounts) extra_doc = text "discounted size =" <+> int discounted_size discounted_size = size - discount small_enough = discounted_size <= ufUseThreshold dflags discount = computeDiscount dflags arg_discounts res_discount arg_infos cont_info where mk_doc some_benefit extra_doc yes_or_no = vcat [ text "arg infos" <+> ppr arg_infos , text "interesting continuation" <+> ppr cont_info , text "some_benefit" <+> ppr some_benefit , text "is exp:" <+> ppr is_exp , text "is work-free:" <+> ppr is_wf , text "guidance" <+> ppr guidance , extra_doc , text "ANSWER =" <+> if yes_or_no then text "YES" else text "NO"] str = "Considering inlining: " ++ showSDocDump dflags (ppr id) n_val_args = length arg_infos -- some_benefit is used when the RHS is small enough -- and the call has enough (or too many) value -- arguments (ie n_val_args >= arity). But there must -- be *something* interesting about some argument, or the -- result context, to make it worth inlining calc_some_benefit :: Arity -> Bool -- The Arity is the number of args -- expected by the unfolding calc_some_benefit uf_arity | not saturated = interesting_args -- Under-saturated -- Note [Unsaturated applications] | otherwise = interesting_args -- Saturated or over-saturated || interesting_call where saturated = n_val_args >= uf_arity over_saturated = n_val_args > uf_arity interesting_args = any nonTriv arg_infos -- NB: (any nonTriv arg_infos) looks at the -- over-saturated args too which is "wrong"; -- but if over-saturated we inline anyway. interesting_call | over_saturated = True | otherwise = case cont_info of CaseCtxt -> not (lone_variable && is_wf) -- Note [Lone variables] ValAppCtxt -> True -- Note [Cast then apply] RuleArgCtxt -> uf_arity > 0 -- See Note [Unfold info lazy contexts] DiscArgCtxt -> uf_arity > 0 -- RhsCtxt -> uf_arity > 0 -- _ -> not is_top && uf_arity > 0 -- Note [Nested functions] -- Note [Inlining in ArgCtxt] {- Note [Unfold into lazy contexts], Note [RHS of lets] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ When the call is the argument of a function with a RULE, or the RHS of a let, we are a little bit keener to inline. For example f y = (y,y,y) g y = let x = f y in ...(case x of (a,b,c) -> ...) ... We'd inline 'f' if the call was in a case context, and it kind-of-is, only we can't see it. Also x = f v could be expensive whereas x = case v of (a,b) -> a is patently cheap and may allow more eta expansion. So we treat the RHS of a let as not-totally-boring. Note [Unsaturated applications] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ When a call is not saturated, we *still* inline if one of the arguments has interesting structure. That's sometimes very important. A good example is the Ord instance for Bool in Base: Rec { $fOrdBool =GHC.Classes.D:Ord @ Bool ... $cmin_ajX $cmin_ajX [Occ=LoopBreaker] :: Bool -> Bool -> Bool $cmin_ajX = GHC.Classes.$dmmin @ Bool $fOrdBool } But the defn of GHC.Classes.$dmmin is: $dmmin :: forall a. GHC.Classes.Ord a => a -> a -> a {- Arity: 3, HasNoCafRefs, Strictness: SLL, Unfolding: (\ @ a $dOrd :: GHC.Classes.Ord a x :: a y :: a -> case @ a GHC.Classes.<= @ a $dOrd x y of wild { GHC.Types.False -> y GHC.Types.True -> x }) -} We *really* want to inline $dmmin, even though it has arity 3, in order to unravel the recursion. Note [Things to watch] ~~~~~~~~~~~~~~~~~~~~~~ * { y = I# 3; x = y `cast` co; ...case (x `cast` co) of ... } Assume x is exported, so not inlined unconditionally. Then we want x to inline unconditionally; no reason for it not to, and doing so avoids an indirection. * { x = I# 3; ....f x.... } Make sure that x does not inline unconditionally! Lest we get extra allocation. Note [Inlining an InlineRule] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ An InlineRules is used for (a) programmer INLINE pragmas (b) inlinings from worker/wrapper For (a) the RHS may be large, and our contract is that we *only* inline when the function is applied to all the arguments on the LHS of the source-code defn. (The uf_arity in the rule.) However for worker/wrapper it may be worth inlining even if the arity is not satisfied (as we do in the CoreUnfolding case) so we don't require saturation. Note [Nested functions] ~~~~~~~~~~~~~~~~~~~~~~~ If a function has a nested defn we also record some-benefit, on the grounds that we are often able to eliminate the binding, and hence the allocation, for the function altogether; this is good for join points. But this only makes sense for *functions*; inlining a constructor doesn't help allocation unless the result is scrutinised. UNLESS the constructor occurs just once, albeit possibly in multiple case branches. Then inlining it doesn't increase allocation, but it does increase the chance that the constructor won't be allocated at all in the branches that don't use it. Note [Cast then apply] ~~~~~~~~~~~~~~~~~~~~~~ Consider myIndex = __inline_me ( (/\a. <blah>) |> co ) co :: (forall a. a -> a) ~ (forall a. T a) ... /\a.\x. case ((myIndex a) |> sym co) x of { ... } ... We need to inline myIndex to unravel this; but the actual call (myIndex a) has no value arguments. The ValAppCtxt gives it enough incentive to inline. Note [Inlining in ArgCtxt] ~~~~~~~~~~~~~~~~~~~~~~~~~~ The condition (arity > 0) here is very important, because otherwise we end up inlining top-level stuff into useless places; eg x = I# 3# f = \y. g x This can make a very big difference: it adds 16% to nofib 'integer' allocs, and 20% to 'power'. At one stage I replaced this condition by 'True' (leading to the above slow-down). The motivation was test eyeball/inline1.hs; but that seems to work ok now. NOTE: arguably, we should inline in ArgCtxt only if the result of the call is at least CONLIKE. At least for the cases where we use ArgCtxt for the RHS of a 'let', we only profit from the inlining if we get a CONLIKE thing (modulo lets). Note [Lone variables] See also Note [Interaction of exprIsWorkFree and lone variables] ~~~~~~~~~~~~~~~~~~~~~ which appears below The "lone-variable" case is important. I spent ages messing about with unsatisfactory varaints, but this is nice. The idea is that if a variable appears all alone as an arg of lazy fn, or rhs BoringCtxt as scrutinee of a case CaseCtxt as arg of a fn ArgCtxt AND it is bound to a cheap expression then we should not inline it (unless there is some other reason, e.g. is is the sole occurrence). That is what is happening at the use of 'lone_variable' in 'interesting_call'. Why? At least in the case-scrutinee situation, turning let x = (a,b) in case x of y -> ... into let x = (a,b) in case (a,b) of y -> ... and thence to let x = (a,b) in let y = (a,b) in ... is bad if the binding for x will remain. Another example: I discovered that strings were getting inlined straight back into applications of 'error' because the latter is strict. s = "foo" f = \x -> ...(error s)... Fundamentally such contexts should not encourage inlining because the context can ``see'' the unfolding of the variable (e.g. case or a RULE) so there's no gain. If the thing is bound to a value. However, watch out: * Consider this: foo = _inline_ (\n. [n]) bar = _inline_ (foo 20) baz = \n. case bar of { (m:_) -> m + n } Here we really want to inline 'bar' so that we can inline 'foo' and the whole thing unravels as it should obviously do. This is important: in the NDP project, 'bar' generates a closure data structure rather than a list. So the non-inlining of lone_variables should only apply if the unfolding is regarded as cheap; because that is when exprIsConApp_maybe looks through the unfolding. Hence the "&& is_wf" in the InlineRule branch. * Even a type application or coercion isn't a lone variable. Consider case $fMonadST @ RealWorld of { :DMonad a b c -> c } We had better inline that sucker! The case won't see through it. For now, I'm treating treating a variable applied to types in a *lazy* context "lone". The motivating example was f = /\a. \x. BIG g = /\a. \y. h (f a) There's no advantage in inlining f here, and perhaps a significant disadvantage. Hence some_val_args in the Stop case Note [Interaction of exprIsWorkFree and lone variables] ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ The lone-variable test says "don't inline if a case expression scrutines a lone variable whose unfolding is cheap". It's very important that, under these circumstances, exprIsConApp_maybe can spot a constructor application. So, for example, we don't consider let x = e in (x,x) to be cheap, and that's good because exprIsConApp_maybe doesn't think that expression is a constructor application. In the 'not (lone_variable && is_wf)' test, I used to test is_value rather than is_wf, which was utterly wrong, because the above expression responds True to exprIsHNF, which is what sets is_value. This kind of thing can occur if you have {-# INLINE foo #-} foo = let x = e in (x,x) which Roman did. -} computeDiscount :: DynFlags -> [Int] -> Int -> [ArgSummary] -> CallCtxt -> Int computeDiscount dflags arg_discounts res_discount arg_infos cont_info -- We multiple the raw discounts (args_discount and result_discount) -- ty opt_UnfoldingKeenessFactor because the former have to do with -- *size* whereas the discounts imply that there's some extra -- *efficiency* to be gained (e.g. beta reductions, case reductions) -- by inlining. = 10 -- Discount of 10 because the result replaces the call -- so we count 10 for the function itself + 10 * length actual_arg_discounts -- Discount of 10 for each arg supplied, -- because the result replaces the call + round (ufKeenessFactor dflags * fromIntegral (total_arg_discount + res_discount')) where actual_arg_discounts = zipWith mk_arg_discount arg_discounts arg_infos total_arg_discount = sum actual_arg_discounts mk_arg_discount _ TrivArg = 0 mk_arg_discount _ NonTrivArg = 10 mk_arg_discount discount ValueArg = discount res_discount' | LT <- arg_discounts `compareLength` arg_infos = res_discount -- Over-saturated | otherwise = case cont_info of BoringCtxt -> 0 CaseCtxt -> res_discount -- Presumably a constructor ValAppCtxt -> res_discount -- Presumably a function _ -> 40 `min` res_discount -- ToDo: this 40 `min` res_discount doesn't seem right -- for DiscArgCtxt it shouldn't matter because the function will -- get the arg discount for any non-triv arg -- for RuleArgCtxt we do want to be keener to inline; but not only -- constructor results -- for RhsCtxt I suppose that exposing a data con is good in general -- And 40 seems very arbitrary -- -- res_discount can be very large when a function returns -- constructors; but we only want to invoke that large discount -- when there's a case continuation. -- Otherwise we, rather arbitrarily, threshold it. Yuk. -- But we want to aovid inlining large functions that return -- constructors into contexts that are simply "interesting"
alexander-at-github/eta
compiler/ETA/Core/CoreUnfold.hs
bsd-3-clause
58,307
0
19
16,859
6,694
3,556
3,138
499
32
f = foo (\x -> \y -> x x y y)
mpickering/hlint-refactor
tests/examples/Lambda18.hs
bsd-3-clause
29
1
9
10
33
15
18
1
1
<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd"> <helpset version="2.0" xml:lang="el-GR"> <title>Image Locaiton and Privacy Scanner | ZAP Extension</title> <maps> <homeID>top</homeID> <mapref location="map.jhm"/> </maps> <view> <name>TOC</name> <label>Contents</label> <type>org.zaproxy.zap.extension.help.ZapTocView</type> <data>toc.xml</data> </view> <view> <name>Index</name> <label>Index</label> <type>javax.help.IndexView</type> <data>index.xml</data> </view> <view> <name>Search</name> <label>Search</label> <type>javax.help.SearchView</type> <data engine="com.sun.java.help.search.DefaultSearchEngine"> JavaHelpSearch </data> </view> <view> <name>Favorites</name> <label>Favorites</label> <type>javax.help.FavoritesView</type> </view> </helpset>
kingthorin/zap-extensions
addOns/imagelocationscanner/src/main/javahelp/org/zaproxy/zap/extension/imagelocationscanner/resources/help_el_GR/helpset_el_GR.hs
apache-2.0
995
104
29
162
429
215
214
-1
-1
import Graphics.Rendering.OpenGL import Graphics.UI.GLUT myPoints :: [(GLfloat,GLfloat,GLfloat)] myPoints = map (\k -> (sin(2*pi*k/12),cos(2*pi*k/12),0.0)) [1..12] main = do (progname, _) <- getArgsAndInitialize createWindow "Hello World" displayCallback $= display mainLoop display = do clear [ColorBuffer] renderPrimitive Points $ mapM_ (\(x, y, z)->vertex$Vertex3 x y z) myPoints flush
erantapaa/haskell-platform
hptool/os-extras/win/test/glut.hs
bsd-3-clause
407
2
13
63
206
103
103
13
1
{-# OPTIONS_GHC -fno-warn-redundant-constraints #-} {-# LANGUAGE TypeFamilies, UndecidableInstances #-} module T2448 where import Data.Kind (Type) -- Demonstrates a bug in propagating type equality constraints class VectorSpace v where type Scalar v :: Type class VectorSpace v => InnerSpace v instance (VectorSpace u,VectorSpace v, Scalar u ~ Scalar v) => VectorSpace (u,v) where type Scalar (u,v) = Scalar u instance (InnerSpace u,InnerSpace v, Scalar u ~ Scalar v) => InnerSpace (u,v)
sdiehl/ghc
testsuite/tests/indexed-types/should_compile/T2448.hs
bsd-3-clause
505
0
8
87
150
80
70
-1
-1
-- Test for trac ticket #1287; ghc 6.6 and 6.6.1 panicked on this module ShouldCompile where {-# SPECIALIZE delta' :: Num b => Int -> Int -> b -> b -> b #-} delta' :: Eq a => a -> a -> b -> b -> b delta' x y e f = if (x==y) then f else e {-# SPECIALIZE delta :: Num b => Int -> Int -> b #-} delta :: (Eq a, Num b) => a -> a -> b delta x y = delta' x y 0 1
spacekitteh/smcghc
testsuite/tests/simplCore/should_compile/simpl016.hs
bsd-3-clause
360
0
9
97
111
60
51
7
2
{-# LANGUAGE OverloadedStrings #-} module Main where import Web.Spock.Safe import Data.Aeson (object, (.=)) import Data.Monoid ((<>)) import Data.Maybe (fromMaybe) import Helpers (breakTextByDot) import System.Environment (lookupEnv) import qualified FizzBuzz as FB main :: IO () main = do maybePort <- lookupEnv "PORT" let port = read $ fromMaybe "3000" maybePort runSpock port $ spockT id $ do get "/" $ redirect "3,Fizz;5,Buzz" get ("/" <//> var) $ \ftype -> do case breakTextByDot ftype of (t, ".json") -> json $ object ["answer" .= FB.fizzerBuzzer t (1, 100)] (t, _) -> text $ FB.fizzerBuzzer t (1, 100)
ifo/fbaas
src/Main.hs
isc
651
0
22
136
245
132
113
19
2
{-# LANGUAGE OverloadedStrings #-} module Commands ( respond ) where import qualified Data.ByteString.Char8 as B import Data.Maybe (listToMaybe) import Data.UUID (fromString) import Data.UUID.V4 (nextRandom) import System.IO (Handle) import Control.Applicative ((<$>)) import Control.Monad.IO.Class import Control.Monad.State import Control.Monad.Except import Control.Monad.Reader import Messages import Game import Actions respond :: Message -> Handle -> GameState -> IO GameState respond message handle state = snd <$> runStateT (runReaderT (cmd (command message) handle) message) state cmd :: B.ByteString -> Handle -> ReaderT Message (StateT GameState IO) () cmd "action" = \handle -> do message <- ask if length (mParts message) < 2 then invalid handle else get >>= liftIO . runAction handle message >>= put cmd "robot_stopped" = \handle -> do message <- ask if length (mParts message) /= 1 then invalid handle else do let idPart = head $ mParts message modify . changeRobot idPart $ \r -> r { status = Off } cmd _ = invalid complain :: B.ByteString -> Handle -> ReaderT Message (StateT GameState IO) () complain err handle = undefined invalid :: Handle -> ReaderT Message (StateT GameState IO) () invalid handle = liftIO . (flip sendInvalid handle) =<< ask
shak-mar/botstrats
server/Commands.hs
mit
1,322
0
15
248
455
242
213
36
3
{-# LANGUAGE TemplateHaskell #-} -- このソースコードは、次の記事を信頼して書かれました. -- http://osecpu.osask.jp/wiki/?page0104 -- http://osecpu.osask.jp/wiki/?page0092 -- 変数のレジスタ割当は,レキシカルスコープのように,ある時点で使用されている変数を管理すれば良い. -- ラベルの割当は,プログラム中でユニークであるべきなので,コンパイル中全体で管理する. module ChageComp where import Data.Word import Data.List (genericLength, tails) import Control.Monad.State import Control.Monad -- Lensなる便利なライブラリを使うことにした import Control.Lens import Control.Applicative import Inst import AST import Type import Typing import qualified IR -- 変数とレジスタの割当関係を管理する. data Frame = Frame { _intVars :: [(Var, Reg)] , _ptrVars :: [(Var, PReg)] , _intRest :: [Reg] , _ptrRest :: [PReg] } makeLenses ''Frame emptyFrame :: Frame emptyFrame = Frame [] [] [Reg 0 .. Reg 0x27] [PReg 0 .. PReg 0x27] -- コンパイル中に保持しておくべき情報.内部的に用いられるのみである. data CompileState = CS { _labels :: Word32, -- 導入したラベルの数 _frame :: Frame } makeLenses ''CompileState -- 変数として確保できるレジスタの数の上限. [R00 ~ R27] localRegisterCount = 0x27 -- テンポラリレジスタとして.[R28 ~ R2B] tmp1 = Reg 0x3A tmp2 = Reg 0x3B -- 32bit指定のやつ. spec32 = BitSpec 32 -- ポインタレジスタを使うとして, p3f = PReg 0x3F p30 = PReg 0x30 p3e = PReg 0x3E p2f = PReg 0x2F -- ラベルには二種類ある.ジャンプするためのラベルと,読み書きするためのラベルだ. jumpOnly = LabelOpt 0 readWrite = LabelOpt 1 apiList = [("api_drawPoint", (Imm 0x0002, [Reg 0x31 .. Reg 0x34])), ("api_drawLine", (Imm 0x0003, [Reg 0x31 .. Reg 0x36])), ("api_drawRect", (Imm 0x0004, [Reg 0x31 .. Reg 0x36])), ("api_drawOval", (Imm 0x0005, [Reg 0x31 .. Reg 0x36])), ("api_exit", (Imm 0x0008, [Reg 0x31 .. Reg 0x31])), ("api_sleep", (Imm 0x0009, [Reg 0x31 .. Reg 0x32])), ("api_openWin", (Imm 0x0010, [Reg 0x31 .. Reg 0x34])) ] -- スコープのネストした中でモナドを走らせる.要は,いったん戻してるだけ. extendScope :: State CompileState a -> State CompileState a extendScope m = do fr <- use frame ret <- m frame .= fr return ret -- 整数の変数を,シンボルテーブルから引っ張ってくる. lookupVar :: Var -> State CompileState Reg lookupVar var@(Var s) = do vars <- use (frame.intVars) case lookup var vars of Nothing -> error $ "undefined variable: " ++ s Just rg -> return rg -- ポインタ変数を,シンボルテーブルから引っ張ってくる. lookupPtr :: Var -> State CompileState PReg lookupPtr ptr@(Var s) = do vars <- use (frame.ptrVars) case lookup ptr vars of Nothing -> error $ "undefined pointer variable: " ++ s Just pr -> return pr alreadyDefinedVar :: Var -> State CompileState Bool alreadyDefinedVar v = do vars <- use (frame.intVars) return $ case lookup v vars of Nothing -> False Just _ -> True alreadyDefinedPtr :: Var -> State CompileState Bool alreadyDefinedPtr v = do vars <- use (frame.ptrVars) return $ case lookup v vars of Nothing -> False Just _ -> True defineVar var@(Var v) = do defd <- alreadyDefinedVar var when defd (error $ "already defined var: " ++ v) reg <- use (frame.intRest) <&> head frame.intVars %= ((var, reg) :) frame.intRest %= tail return $ reg definePtr ptr@(Var v) = do defd <- alreadyDefinedPtr ptr when defd (error $ "already defined ptr: " ++ v) reg <- use (frame.ptrRest) <&> head frame.ptrVars %= ((ptr, reg) :) frame.ptrRest %= tail return $ reg -- releaseVar var@(Var v) = do -- frame.intVars %= filter ((v ==) . fst) genLabel :: State CompileState Label genLabel = do lb <- use labels labels += 1 return $ Label lb arith And = AND arith Xor = XOR arith Or = OR arith Add = ADD arith Mul = MUL comp Cmpe = CMPE comp Cmpne = CMPNE comp Cmpl = CMPL comp Cmple = CMPLE comp Cmpg = CMPG comp Cmpge = CMPGE compile :: [IR.IR] -> Program compile is = Program $ evalState (compile' is) (CS 0 emptyFrame) where compile' :: [IR.IR] -> State CompileState [Inst] compile' irs = concat <$> zipWithM compileIR irs (map concat $ tails $ map IR.varRefs irs) compileIR :: IR.IR -> [Var] -> State CompileState [Inst] compileIR ir livingVars = case ir of IR.If var csq alt -> do cond <- lookupVar var c1 <- extendScope (compile' csq) c2 <- extendScope (compile' alt) ifLabel <- genLabel endLabel <- genLabel return $ [LIMM spec32 tmp2 (Imm 1), XOR spec32 tmp1 cond tmp2, CND tmp1, PLIMM p3f ifLabel] ++ c1 ++ [PLIMM p3f endLabel, LB jumpOnly ifLabel] ++ c2 ++ [LB jumpOnly endLabel] IR.While var cond body -> do c1 <- compile' cond cv <- lookupVar var c2 <- extendScope (compile' body) startLabel <- genLabel endLabel <- genLabel return $ [LB jumpOnly startLabel] ++ c1 ++ [LIMM spec32 tmp2 (Imm 1), XOR spec32 tmp1 cv tmp2, CND tmp1, PLIMM p3f endLabel] ++ c2 ++ [PLIMM p3f startLabel, LB jumpOnly endLabel] IR.Declare var S32Int val -> do defineVar var r <- lookupVar var v <- lookupVar val return [OR spec32 r v v] IR.Declare var (Pointer _) val -> do definePtr var p <- lookupPtr var v <- lookupPtr val return [PCP p v] IR.Arith op var e1 e2 -> do defineVar var r <- lookupVar var v1 <- lookupVar e1 v2 <- lookupVar e2 return [arith op spec32 r v1 v2] IR.Comp op var e1 e2 -> do defineVar var r <- lookupVar var v1 <- lookupVar e1 v2 <- lookupVar e2 return [comp op spec32 spec32 r v1 v2] IR.ConstS32Int var int -> do defineVar var r <- lookupVar var return [LIMM spec32 r (Imm (fromInteger (toInteger int)))] IR.Assign var S32Int val -> do r <- lookupVar var v <- lookupVar val return [OR spec32 r v v] IR.Assign var (Pointer _) val -> do p <- lookupPtr var v <- lookupPtr val return [PCP p v] IR.Load var ptr idx -> do defineVar var v <- lookupVar var p <- lookupPtr ptr i <- lookupVar idx return [PADD spec32 p3e p i, LMEM0 spec32 v p3e] IR.Store ptr idx val -> do p <- lookupPtr ptr i <- lookupVar idx v <- lookupVar val return [PADD spec32 p3e p i, SMEM0 spec32 v p3e] IR.Call name vars -> case lookup name apiList of Nothing -> error $ "function not found: " ++ name Just (inst, regs) -> do args <- mapM lookupVar vars let move = zipWith (\d s -> OR spec32 d s s) regs args lb <- genLabel return $ move ++ [LIMM spec32 (Reg 0x30) inst, PLIMM p30 lb, PCP p3f p2f, LB readWrite lb] IR.Data name dat -> do definePtr name p <- lookupPtr name lb <- genLabel return $ [PLIMM p lb, LB readWrite lb, DATA dat] IR.DebugStop -> return [BREAK] -- <* mapM_ (\v -> unless (v `elem` livingVars) re) IR.varRefs ir test1 = AST [ Declare (Var "x") S32Int (Arith () Add (Arith () Xor (ConstS32Int () 3) (ConstS32Int () 4)) (ConstS32Int () 5))] test2 = AST [ Data (Var "d") [1, 2, 3], Declare (Var "y") S32Int (Load () (GetVar () (Var "d")) (ConstS32Int () 1)), Store (GetVar () (Var "d")) (ConstS32Int () 3) (ConstS32Int () 4) ] test3 = AST [ Data (Var "d") [1, 2, 3], Declare (Var "y") (Pointer S32Int) (GetVar () (Var "d")) ] test4 = AST [ Declare (Var "i") S32Int (ConstS32Int () 0), While (GetVar () (Var "i")) (AST []) ] test5 = AST [ If (Comp () Cmpe (ConstS32Int () 3) (ConstS32Int () 4)) (AST []) (AST []) ] process = compile . IR.normalize . typing hoge = process
ryna4c2e/chage
ChageComp.hs
mit
9,595
0
21
3,406
2,929
1,441
1,488
-1
-1
module Handler.Euler where import Import import Yesod.Form.Bootstrap3 getEulerR :: Handler Html getEulerR = error "Not yet implemented: getEulerR" postEulerR :: Handler Html postEulerR = error "Not yet implemented: postEulerR" getEulerSolutionR :: ProblemNumber -> Handler Html getEulerSolutionR = error "Not yet implemented: getEulerSolutionR" postEulerSolutionR :: ProblemNumber -> Handler Html postEulerSolutionR num = error "Not yet implemented: postEulerSolutionR" solutionForm :: UserId -> Form EulerSolution solutionForm userId = renderBootstrap3 BootstrapBasicForm $ EulerSolution <$> pure userId <*> (ProblemNumber <$> areq intField "Problem Number" Nothing) <*> areq intField "Solution" Nothing <*> (unTextarea <$> areq textareaField "Markdown" Nothing) <*> pure Nothing
terrelln/terrelln.me
Handler/Euler.hs
mit
807
0
11
121
178
89
89
18
1
{-# LANGUAGE CPP #-} module Test.Mockery.ActionSpec (spec) where import Test.Hspec import Control.Monad import Test.HUnit.Lang import Test.Mockery.Action #if MIN_VERSION_HUnit(1,4,0) hUnitFailure :: String -> HUnitFailure -> Bool hUnitFailure actual (HUnitFailure _ reason) = case reason of Reason expected -> actual == expected _ -> False #else hUnitFailure :: String -> HUnitFailure -> Bool hUnitFailure actual (HUnitFailure _ expected) = actual == expected #endif spec :: Spec spec = do describe "dummy" $ do it "fails" $ do (dummy "test" :: Int -> Int -> IO Int) 23 42 `shouldThrow` hUnitFailure "Unexpected call to dummy action: test" describe "stub" $ do context "with one parameter" $ do context "when receiving specified parameters" $ do it "returns specified value" $ do stub ("foo", return "r") "foo" `shouldReturn` "r" context "when receiving unexpected parameters" $ do it "throws an exception" $ do stub ("foo", return "r") "bar" `shouldThrow` (hUnitFailure . unlines) [ "Unexected parameter to stubbed action!" , "expected: " ++ show "foo" , " but got: " ++ show "bar" ] context "with two parameters" $ do context "when receiving specified parameters" $ do it "returns specified value" $ do stub ("foo", "bar", return "r") "foo" "bar" `shouldReturn` "r" context "when receiving unexpected parameters" $ do it "throws an exception" $ do stub ("foo", "bar", return "r") "23" "42" `shouldThrow` (hUnitFailure . unlines) [ "Unexected parameters to stubbed action!" , "expected: " ++ show ("foo", "bar") , " but got: " ++ show ("23", "42") ] context "with three parameters" $ do context "when receiving specified parameters" $ do it "returns specified value" $ do stub ("foo", "bar", "baz", return "r") "foo" "bar" "baz" `shouldReturn` "r" context "when receiving unexpected parameters" $ do it "throws an exception" $ do stub ("foo", "bar", "baz", return "r") "23" "42" "65" `shouldThrow` (hUnitFailure . unlines) [ "Unexected parameters to stubbed action!" , "expected: " ++ show ("foo", "bar", "baz") , " but got: " ++ show ("23", "42", "65") ] context "when used with lists" $ do context "with two parameters" $ do context "when receiving specified parameters" $ do it "returns specified value" $ do stub [("foo", "bar", return "r"), ("foo", "baz", return "_")] "foo" "bar" `shouldReturn` "r" context "when receiving unexpected parameters" $ do it "throws an exception" $ do stub [(10, 20, return ()), (23, 42, return ())] (23 :: Int) (65 :: Int) `shouldThrow` (hUnitFailure . unlines) [ "Unexected parameters to stubbed action!" , "expected one of: (10,20), (23,42)" , " but got: (23,65)" ] describe "withMock" $ do let withMockSpec stubbedAction call = do context "when action is called once" $ do it "passes" $ do withMock stubbedAction $ \action -> do call action `shouldReturn` "r" context "when action is called multiple times" $ do it "fails" $ do withMock stubbedAction $ \action -> do replicateM_ 10 (call action) `shouldThrow` hUnitFailure "Expected to be called once, but it was called 10 times instead!" context "when action is not called" $ do it "fails" $ do withMock stubbedAction $ \_ -> do return () `shouldThrow` hUnitFailure "Expected to be called once, but it was called 0 times instead!" context "with one parameter" $ do let stubbedAction = stub ("foo", return "r") call action = action "foo" withMockSpec stubbedAction call context "with two parameters" $ do let stubbedAction = stub ("foo", "bar", return "r") call action = action "foo" "bar" withMockSpec stubbedAction call context "with three parameters" $ do let stubbedAction = stub ("foo", "bar", "baz", return "r") call action = action "foo" "bar" "baz" withMockSpec stubbedAction call describe "mockChain" $ do let actions = replicate 2 $ stub ("foo", "bar", return "r") context "when mock is called the specified number of times" $ do it "passes" $ do mockChain actions $ \mock -> do replicateM_ 2 (mock "foo" "bar") context "when mock is called too often" $ do it "fails" $ do mockChain actions $ \mock -> do replicateM_ 3 (mock "foo" "bar") `shouldThrow` hUnitFailure "Expected to be called only 2 times, but it received an additional call!" context "when mock is not called often enough" $ do it "fails" $ do mockChain actions $ \mock -> do replicateM_ 1 (mock "foo" "bar") `shouldThrow` hUnitFailure "Expected to be called 2 times, but it was called 1 time instead!"
robbinch/tinc
test/Test/Mockery/ActionSpec.hs
mit
5,251
0
29
1,590
1,370
657
713
101
1
import Data.Char import Data.List import System.Environment( getArgs ) import System.Random import System.Console.GetOpt main = do args <- getArgs let key = head args randomNote <- randomRIO (0, 6) randomSuff <- randomRIO (0, 6) -- prints random key in scale putStrLn $ (deriveScale key !! randomNote) -- prints random suffix putStrLn $ (suffix !! randomSuff) -- sharp chromatic scale, for mapping sharps = ["C", "C#", "D", "D#", "E", "F", "F#", "G", "G#", "A", "A#", "B"] -- flat chromatic scale, for mapping flats = ["C", "Db", "D", "Eb", "E", "F", "Gb", "G", "Ab", "A", "Bb", "B"] suffix = ["", "6", "6\n4", "7", "6\n5", "4\n3", "4\n2"] -- isSharps a returns True if key a is a sharp key, -- and False if key a is a flat key isSharps :: String -> Bool -- edge case: the only element in sharps that is a flat key isSharps a | a == "F" = False | (map toUpper a) `elem` sharps = True isSharps a = False {- deriveOffset a returns the offset from the key of C e.g.) deriveOffset D returns 2, because D is offset from C by 2 deriveOffset :: String -> Int these are our outside cases, not appearing in sharps (we only need to determine index here): Db, db, Eb, eb, Gb, gb, Ab, ab, Bb, bb we match both sides toLower equivalent e.g.) map toLower Bb == "bb" because matching toUpper results in B not matching -} deriveOffset key | map toLower key == "db" = 1 | map toLower key == "eb" = 3 | map toLower key == "gb" = 6 | map toLower key == "ab" = 8 | map toLower key == "bb" = 10 deriveOffset key = head (elemIndices (map toUpper key) sharps) -- maps sharp or flat scale over scale intervals from derivedModality deriveScale :: String -> [String] deriveScale a | isSharps a = map (sharps !!) (deriveModality a) deriveScale a = map (flats !!) (deriveModality a) -- creates scale intervals (major or minor), using offset amount deriveModality :: String -> [Int] -- key is major deriveModality a | isUpper (head a) = [(0 + offset) `mod` 12, (2 + offset) `mod` 12, (4 + offset) `mod` 12, (5 + offset) `mod` 12, (7 + offset) `mod` 12, (9 + offset) `mod` 12, (11 + offset) `mod` 12] where offset = deriveOffset a -- key is minor deriveModality a = [(0 + offset) `mod` 12, (2 + offset) `mod` 12, (3 + offset) `mod` 12, (5 + offset) `mod` 12, (7 + offset) `mod` 12, (8 + offset) `mod` 12, (10 + offset) `mod` 12] where offset = deriveOffset a
rebennett/figbass
figbass.hs
mit
2,508
5
10
617
816
442
374
31
1
module Main where import Protolude import Lib main :: IO () main = someFunc
Archimidis/md2html
app/Main.hs
mit
78
0
6
16
25
15
10
5
1
{-# LANGUAGE NoMonomorphismRestriction #-} {-# LANGUAGE TupleSections #-} {-# LANGUAGE FlexibleContexts #-} {-# LANGUAGE ScopedTypeVariables #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE Rank2Types #-} {-# LANGUAGE TemplateHaskell #-} {-# LANGUAGE LambdaCase #-} module StorePackage where import qualified Codec.Archive.Tar as Tar import Control.Applicative import Control.Monad import Control.Monad.IO.Class import Control.Monad.Trans.Resource import Control.Exception.Lifted import Data.Aeson ((.:), (.=)) import qualified Data.Aeson as JSON import qualified Data.ByteString.Char8 as S import qualified Data.ByteString.Lazy.Char8 as L import Data.Conduit import qualified Data.Conduit.List as CL import qualified Data.Conduit.Zlib as ZLib import Data.Function (fix, on) import qualified Data.HashMap.Strict as H import Data.List import Data.Maybe import qualified Data.Text as T import qualified Data.Text.Encoding as T import qualified Data.Vector as V import qualified Distribution.Package as D import qualified Distribution.PackageDescription as D import qualified Distribution.PackageDescription.Configuration as D import qualified Distribution.PackageDescription.Parse as D import qualified Distribution.Text as D import Network.HTTP.Client.TLS import Network.HTTP.Conduit import System.Environment import Numeric import Data.Time import System.Locale import System.IO untar :: MonadThrow m => Conduit S.ByteString m Tar.Entry untar = do s <- CL.consume loop $ Tar.read (L.fromChunks s) where loop (Tar.Next e es) = yield e >> loop es loop Tar.Done = return () loop (Tar.Fail e) = throwM e isCabalFile :: Tar.Entry -> Bool isCabalFile e = ".cabal" `isSuffixOf` Tar.entryPath e normalFileContent :: Tar.EntryContent -> Maybe L.ByteString normalFileContent (Tar.NormalFile c _) = Just c normalFileContent _ = Nothing resultMaybe :: D.ParseResult a -> Maybe a resultMaybe (D.ParseOk _ a) = Just a resultMaybe _ = Nothing entryToPackageDescription :: (MonadBaseControl IO m, MonadIO m) => Request -> Manager -> Conduit Tar.Entry m (D.PackageDescription, UTCTime) entryToPackageDescription req mgr = CL.filter isCabalFile =$ CL.mapMaybe (normalFileContent . Tar.entryContent) =$ CL.mapMaybe (resultMaybe . D.parsePackageDescription . L.unpack) =$ CL.map D.flattenPackageDescription =$ CL.groupBy ((==) `on` (D.pkgName . D.package)) =$ CL.map (maximumBy (compare `on` (D.pkgVersion . D.package))) =$ CL.mapM (\pd -> (pd,) <$> getLastUploaded (D.display . D.pkgName $ D.package pd) req mgr) pdToValue :: D.PackageDescription -> UTCTime -> Maybe [T.Text] -> JSON.Value pdToValue p upl dpr = JSON.object [ "name" .= name , "version" .= (D.display . D.pkgVersion . D.package) p , "license" .= (D.display . D.license) p , "copyright" .= D.copyright p , "maintainer" .= D.maintainer p , "author" .= D.author p , "stability" .= D.stability p , "homepage" .= D.homepage p , "bugReports" .= D.bugReports p , "synopsis" .= D.synopsis p , "description" .= (unescapeHtml . D.description) p , "category" .= (splitCategory . D.category) p , "hasLibrary" .= (isJust . D.library) p , "hasExecutable" .= (not . null . D.executables) p , "executables" .= (map D.exeName . D.executables) p , "deprecated" .= isJust dpr , "inFavourOf" .= maybe [] id dpr , "lastUploaded" .= upl , "ngram" .= JSON.object [ "description" JSON..= (unescapeHtml . D.description) p , "synopsis" JSON..= (unescapeHtml . D.synopsis) p , "name" JSON..= name ] , "raw" .= JSON.object [ "name" .= name , "license" .= (D.display . D.license) p , "category" .= (splitCategory . D.category) p ] ] where splitCategory = map T.strip . T.splitOn "," . T.pack name = (D.display . D.pkgName . D.package) p unescapeHtml :: String -> String unescapeHtml = loop where loop [] = [] loop ('&':cs) = unescape cs loop (c:cs) = c : loop cs entities = H.fromList [("lt", '<'), ("gt", '>'), ("amp", '&'), ("quot", '"')] unescape cs = case break (== ';') cs of (_, []) -> [] ('#':x:hex, _:cs') | x `elem` ['x', 'X'] -> case readHex hex of [(n, [])] -> toEnum n : loop cs' _ -> loop cs' ('#':dec, _:cs') -> case reads dec of [(n, [])] -> toEnum n : loop cs' _ -> loop cs' (ent, _:cs') -> case H.lookup ent entities of Just c -> c : loop cs' Nothing -> loop cs' parseUrl' :: MonadThrow m => String -> m Request parseUrl' s | '@' `notElem` s = parseUrl s parseUrl' s0 = do let (proto, s1) = T.breakOnEnd "://" (T.pack s0) (user, s2) = T.breakOnEnd ":" s1 (pass, s3) = T.breakOnEnd "@" s2 req <- parseUrl . concat $ map T.unpack [proto, s3] return $ applyBasicAuth (T.encodeUtf8 $ T.init user) (T.encodeUtf8 $ T.init pass) req defaultTime :: UTCTime defaultTime = UTCTime (ModifiedJulianDay 0) 0 getLastUploaded :: (MonadIO m, MonadBaseControl IO m) => String -> Request -> Manager -> m UTCTime getLastUploaded pkg req mgr = fmap (maybe defaultTime id . either (\(_ :: SomeException) -> Nothing) id) . try $ parseTime defaultTimeLocale "%a %b %e %X %Z %Y" . L.unpack . responseBody <$> httpLbs req { path = S.pack $ "/package/" ++ pkg ++ "/upload-time" } mgr sinkStoreElasticsearch :: (MonadThrow m, MonadIO m) => Int -> Bool -> H.HashMap T.Text [T.Text] -> Request -> Manager -> Consumer (D.PackageDescription, UTCTime) m () sinkStoreElasticsearch cs progress dpr req' mgr = do let req = req' { method = "POST" } fix $ \loop -> do chunk <- CL.take cs let cmd i = JSON.object ["index" .= JSON.object ["_id" .= (i :: String)]] body = L.unlines . map JSON.encode $ concatMap (\(p,upd) -> let pkg = D.display . D.pkgName $ D.package p in [ cmd pkg , pdToValue p upd (H.lookup (T.pack pkg) dpr) ]) chunk unless (null chunk) $ do _ <- liftIO $ httpLbs req { requestBody = RequestBodyLBS body } mgr when progress . liftIO $ putChar '.' >> hFlush stdout loop newtype Deprecateds = Deprecateds { unDeprecateds :: H.HashMap T.Text [T.Text] } instance JSON.FromJSON Deprecateds where parseJSON (JSON.Array a) = Deprecateds . H.fromList <$> mapM parseElem (mapMaybe fromObject (V.toList a)) where fromObject (JSON.Object o) = Just o fromObject _ = Nothing parseElem o = (,) <$> o .: "deprecated-package" <*> o .: "in-favour-of" parseJSON _ = mzero getDeprecateds :: (MonadThrow m, MonadIO m) => Manager -> m (H.HashMap T.Text [T.Text]) getDeprecateds mgr = do req <- parseUrl "http://hackage.haskell.org/packages/deprecated" res <- httpLbs req { requestHeaders = ("Accept", "application/json") : requestHeaders req } mgr maybe (fail "getDeprecateds: decode json failed.") return $ unDeprecateds <$> (JSON.decode $ responseBody res) storeElasticsearch :: (MonadBaseControl IO m, MonadThrow m, MonadIO m, MonadResource m) => String -> String -> Manager -> m () storeElasticsearch url indexName mgr = do dpr <- getDeprecateds mgr index <- parseUrl "http://hackage.haskell.org/packages/index.tar.gz" src <- http index mgr req <- parseUrl' url responseBody src $$+- ZLib.ungzip =$ untar =$ entryToPackageDescription index mgr =$ sinkStoreElasticsearch 500 True dpr req { path = S.pack $ '/' : indexName ++ "/package/_bulk" } mgr main :: IO () main = withManagerSettings tlsManagerSettings $ \mgr -> liftIO getArgs >>= \case [index, url] -> storeElasticsearch url index mgr _ -> liftIO $ putStrLn "USAGE: main INDEX ELASTICSEARCH_URL"
philopon/find-hackage
src/StorePackage.hs
mit
8,633
0
26
2,552
2,776
1,457
1,319
171
9
-- | -- Types and instances of anonymous records. module Record where import BasePrelude hiding (Proxy) import Data.Functor.Identity import GHC.TypeLits import Foreign.Storable import Foreign.Ptr (plusPtr) import Control.Lens.Basic import qualified Record.TH as TH -- | -- A specialised version of "Data.Proxy.Proxy". -- Defined for compatibility with \"base-4.6\", -- since @Proxy@ was only defined in \"base-4.7\". data FieldName (t :: Symbol) -- | -- Defines a lens to manipulate some value of a type by a type-level name, -- using the string type literal functionality. -- -- Instances are provided for all records and for tuples of arity of up to 24. -- -- Here's how you can use it with tuples: -- -- >trd :: Field "3" a a' v v' => a -> v -- >trd = view (fieldLens (undefined :: FieldName "3")) -- -- The function above will get you the third item of any tuple, which has it. class Field (n :: Symbol) a a' v v' | n a -> v, n a' -> v', n a v' -> a', n a' v -> a where fieldLens :: FieldName n -> Lens a a' v v' instance Field "1" (Identity v1) (Identity v1') v1 v1' where fieldLens = const $ \f -> fmap Identity . f . runIdentity -- Generate the tuple instances of the Field class: return $ do arity <- [2 .. 24] fieldIndex <- [1 .. arity] return $ TH.tupleFieldInstanceDec arity fieldIndex -- | -- A simplified field constraint, -- which excludes the possibility of type-changing updates. type Field' n a v = Field n a a v v -- * Record types and instances ------------------------- -- Generate the record types and instances: return $ do arity <- [1 .. 24] strict <- [False, True] let recordType = TH.recordTypeDec strict arity fieldInstances = do fieldIndex <- [1 .. arity] return $ TH.recordFieldInstanceDec strict arity fieldIndex storableInstance = TH.recordStorableInstanceDec strict arity in recordType : storableInstance : fieldInstances -- * Record construction functions with field names ------------------------- -- Generate the function declarations: return $ concat $ TH.recordConFunDecs <$> [False, True] <*> [1 .. 24]
nikita-volkov/record
library/Record.hs
mit
2,122
0
16
430
423
247
176
-1
-1
module E16 where definition usedArgument unusedArgument = usedArgument {- definition usedArgument = usedArgument -}
pascal-knodel/haskell-craft
Examples/· Errors/E16.hs
mit
141
0
5
39
15
9
6
3
1
test = if isNothing x then (-1.0) else fromJust x
alphaHeavy/hlint
tests/brackets.hs
gpl-2.0
50
1
7
10
28
13
15
1
2
{-# OPTIONS -funbox-strict-fields -O2 #-} module Support.MD5( Hash(), emptyHash, md5,md5file,md5lazy,md5lazyIO, md5show32,md5Bytes,md5String,md5Handle,hashToBytes) where import Control.Monad import Data.Binary import Data.Char import Foreign import Foreign.C import System.IO import qualified Data.ByteString as BS import qualified Data.ByteString.Lazy as LBS import qualified Data.ByteString.Unsafe as BS data Hash = Hash !Word32 !Word32 !Word32 !Word32 deriving(Eq,Ord) md5 :: BS.ByteString -> Hash md5 bs = unsafePerformIO $ allocaBytes 16 $ \digest -> do BS.unsafeUseAsCStringLen bs $ \ (x,y) -> md5Data (castPtr x) (fromIntegral y) digest readDigest digest md5lazy :: LBS.ByteString -> Hash md5lazy lbs = unsafePerformIO $ md5lazyIO lbs md5lazyIO :: LBS.ByteString -> IO Hash md5lazyIO lbs = do allocaBytes (fromIntegral $ get_md5_statesize) $ \msp -> do let ms = MState msp md5_init ms forM_ (LBS.toChunks lbs) $ \bs -> do BS.unsafeUseAsCStringLen bs $ \ (x,y) -> md5_append ms (castPtr x) (fromIntegral y) allocaBytes 16 $ \digest -> do md5_finish ms digest readDigest digest readDigest digest = do w1 <- peekWord32 digest 0 w2 <- peekWord32 digest 4 w3 <- peekWord32 digest 8 w4 <- peekWord32 digest 12 return $ Hash w1 w2 w3 w4 peekWord32 ptr off = do b1 <- peekByteOff ptr off :: IO Word8 b2 <- peekByteOff ptr (off + 1) :: IO Word8 b3 <- peekByteOff ptr (off + 2) :: IO Word8 b4 <- peekByteOff ptr (off + 3) :: IO Word8 let fi = fromIntegral :: Word8 -> Word32 return (fi b1 `shiftL` 24 .|. fi b2 `shiftL` 16 .|. fi b3 `shiftL` 8 .|. fi b4) instance Binary Hash where put (Hash a b c d) = put a >> put b >> put c >> put d get = return Hash `ap` get `ap` get `ap` get `ap` get md5file :: FilePath -> IO Hash md5file fp = md5lazy `fmap` LBS.readFile fp newtype MState = MState (Ptr MState) foreign import ccall unsafe "md5_data" md5Data :: Ptr Word8 -> CInt -> Ptr Word8 -> IO () foreign import ccall unsafe md5_init :: MState -> IO () foreign import ccall unsafe md5_append :: MState -> Ptr Word8 -> CInt -> IO () foreign import ccall unsafe md5_finish :: MState -> Ptr Word8 -> IO () foreign import ccall unsafe get_md5_statesize :: CInt hashToBytes :: Hash -> [Word8] hashToBytes (Hash a b c d) = tb a . tb b . tb c . tb d $ [] where tb :: Word32 -> [Word8] -> [Word8] tb n = showIt 4 n showIt :: Int -> Word32 -> [Word8] -> [Word8] showIt 0 _ r = r showIt i x r = case quotRem x 256 of (y, z) -> let c = fromIntegral z in c `seq` showIt (i-1) y (c:r) md5show32 :: Hash -> String md5show32 hash = f [] (hashToBytes hash) where f cs [] = cs f cs (o1:o2:o3:o4:o5:rest) = f ns rest where i1 = o1 `shiftR` 3 i2 = (o1 `shiftL` 2 .|. o2 `shiftR` 6) .&. 0x1f i3 = o2 `shiftR` 1 .&. 0x1f i4 = (o2 `shiftL` 4 .|. o3 `shiftR` 4) .&. 0x1f i5 = (o3 `shiftL` 1 .|. o4 `shiftR` 7) .&. 0x1f i6 = o4 `shiftR` 2 .&. 0x1f i7 = (o4 `shiftL` 3 .|. o5 `shiftR` 5) .&. 0x1f i8 = o5 .&. 0x1f ns = g i1:g i2:g i3:g i4:g i5:g i6:g i7:g i8:cs g x | x <= 9 = chr (ord '0' + fromIntegral x) | otherwise = chr (ord 'a' + fromIntegral x - 10) f cs ns = reverse (take ((lns * 8 + 4) `div` 5) (f [] (ns ++ replicate (5 - lns) 0))) ++ cs where lns = length ns instance Show Hash where showsPrec _ (Hash a b c d) = showAsHex a . showAsHex b . showAsHex c . showAsHex d showAsHex :: Word32 -> ShowS showAsHex n = showIt 8 n where showIt :: Int -> Word32 -> String -> String showIt 0 _ r = r showIt i x r = case quotRem x 16 of (y, z) -> let c = intToDigit (fromIntegral z) in c `seq` showIt (i-1) y (c:r) emptyHash = Hash 0 0 0 0 md5Bytes :: [Word8] -> Hash md5Bytes bs = unsafePerformIO $ allocaBytes 16 $ \digest -> do withArrayLen bs $ \y x -> md5Data (castPtr x) (fromIntegral y) digest readDigest digest md5String :: String -> Hash md5String ss = md5Bytes (toUTF ss) where -- | Convert Unicode characters to UTF-8. toUTF :: String -> [Word8] toUTF [] = [] toUTF (x:xs) | ord x<=0x007F = (fromIntegral $ ord x):toUTF xs | ord x<=0x07FF = fromIntegral (0xC0 .|. ((ord x `shift` (-6)) .&. 0x1F)): fromIntegral (0x80 .|. (ord x .&. 0x3F)): toUTF xs | otherwise = fromIntegral (0xE0 .|. ((ord x `shift` (-12)) .&. 0x0F)): fromIntegral (0x80 .|. ((ord x `shift` (-6)) .&. 0x3F)): fromIntegral (0x80 .|. (ord x .&. 0x3F)): toUTF xs -- XXX inefficient, don't use it. md5Handle :: Handle -> IO Hash md5Handle h = do hSeek h AbsoluteSeek 0 len <- fromIntegral `liftM` hFileSize h allocaBytes len $ \ptr -> do cnt <- hGetBuf h ptr len unless (cnt == len) $ fail "md5File - read returned too few bytes" hSeek h AbsoluteSeek 0 allocaBytes 16 $ \digest -> do md5Data ptr (fromIntegral len) digest readDigest digest
dec9ue/jhc_copygc
src/Support/MD5.hs
gpl-2.0
5,314
0
19
1,600
2,267
1,152
1,115
-1
-1
{-# LANGUAGE CPP #-} module Hkl.Sixs ( main_sixs ) where import Prelude hiding (concat, head, print) #if __GLASGOW_HASKELL__ < 710 import Control.Applicative ((<$>), (<*>)) #endif import Data.ByteString.Char8 (pack) import Data.Vector.Storable (concat, head) import Control.Exception (bracket) import Control.Monad (forM_) import Numeric.LinearAlgebra (Matrix) import Numeric.Units.Dimensional.Prelude (meter, nano, (*~)) import Pipes (Producer, runEffect, (>->), lift, yield) import Pipes.Prelude (print) import System.FilePath.Posix ((</>)) import Hkl.Types import Hkl.H5 {-# ANN module "HLint: ignore Use camelCase" #-} data DataFrameH5Path = DataFrameH5Path { h5pImage :: DataItem , h5pMu :: DataItem , h5pOmega :: DataItem , h5pDelta :: DataItem , h5pGamma :: DataItem , h5pUB :: DataItem , h5pWavelength :: DataItem , h5pDiffractometerType :: DataItem } deriving (Show) data DataFrameH5 = DataFrameH5 { h5image :: Dataset , h5mu :: Dataset , h5omega :: Dataset , h5delta :: Dataset , h5gamma :: Dataset , h5ub :: Dataset , h5wavelength :: Dataset , h5dtype :: Dataset } data DataFrame = DataFrame { df_n :: Int , df_geometry :: Geometry , df_ub :: Matrix Double } deriving (Show) withDataframeH5 :: File -> DataFrameH5Path -> (DataFrameH5 -> IO r) -> IO r withDataframeH5 h5file dfp = bracket (hkl_h5_open h5file dfp) hkl_h5_close hkl_h5_open :: File -> DataFrameH5Path -> IO DataFrameH5 hkl_h5_open h5file dp = DataFrameH5 <$> openDataset' h5file (h5pImage dp) <*> openDataset' h5file (h5pMu dp) <*> openDataset' h5file (h5pOmega dp) <*> openDataset' h5file (h5pDelta dp) <*> openDataset' h5file (h5pGamma dp) <*> openDataset' h5file (h5pUB dp) <*> openDataset' h5file (h5pWavelength dp) <*> openDataset' h5file (h5pDiffractometerType dp) where openDataset' hid (DataItem name _) = openDataset hid (pack name) Nothing hkl_h5_is_valid :: DataFrameH5 -> IO Bool hkl_h5_is_valid df = do True <- check_ndims (h5mu df) 1 True <- check_ndims (h5omega df) 1 True <- check_ndims (h5delta df) 1 True <- check_ndims (h5gamma df) 1 return True hkl_h5_close :: DataFrameH5 -> IO () hkl_h5_close d = do closeDataset (h5image d) closeDataset (h5mu d) closeDataset (h5omega d) closeDataset (h5delta d) closeDataset (h5gamma d) closeDataset (h5ub d) closeDataset (h5wavelength d) closeDataset (h5dtype d) getDataFrame' :: DataFrameH5 -> Int -> IO DataFrame getDataFrame' d i = do mu <- get_position (h5mu d) i omega <- get_position (h5omega d) i delta <- get_position (h5delta d) i gamma <- get_position (h5gamma d) i wavelength <- get_position (h5wavelength d) 0 ub <- get_ub (h5ub d) let positions = concat [mu, omega, delta, gamma] let source = Source (head wavelength *~ nano meter) return DataFrame { df_n = i , df_geometry = Geometry Uhv source positions Nothing , df_ub = ub } getDataFrame :: DataFrameH5 -> Producer DataFrame IO () getDataFrame d = do (Just n) <- lift $ lenH5Dataspace (h5mu d) forM_ [0..n-1] (\i -> lift (getDataFrame' d i) >>= yield) main_sixs :: IO () main_sixs = do let root = "/nfs/ruche-sixs/sixs-soleil/com-sixs/2015/Shutdown4-5/XpadAu111/" let filename = "align_FLY2_omega_00045.nxs" let dataframe_h5p = DataFrameH5Path { h5pImage = DataItem "com_113934/scan_data/xpad_image" StrictDims , h5pMu = DataItem "com_113934/scan_data/UHV_MU" ExtendDims , h5pOmega = DataItem "com_113934/scan_data/UHV_OMEGA" ExtendDims , h5pDelta = DataItem "com_113934/scan_data/UHV_DELTA" ExtendDims , h5pGamma = DataItem "com_113934/scan_data/UHV_GAMMA" ExtendDims , h5pUB = DataItem "com_113934/SIXS/I14-C-CX2__EX__DIFF-UHV__#1/UB" StrictDims , h5pWavelength = DataItem "com_113934/SIXS/Monochromator/wavelength" StrictDims , h5pDiffractometerType = DataItem "com_113934/SIXS/I14-C-CX2__EX__DIFF-UHV__#1/type" StrictDims } withH5File (root </> filename) $ \h5file -> withDataframeH5 h5file dataframe_h5p $ \dataframe_h5 -> do True <- hkl_h5_is_valid dataframe_h5 runEffect $ getDataFrame dataframe_h5 >-> print
picca/hkl
contrib/haskell/src/Hkl/Sixs.hs
gpl-3.0
5,041
0
15
1,655
1,283
668
615
105
1
module Language.Subleq.Assembly.Prim where import Data.Maybe import Data.Set (Set) import qualified Data.Set as S import Data.Map (Map) import qualified Data.Map as M import Text.Printf import Data.List import Data.Bits type Id = String type Location = String type Substitution = Map Id Expr data Expr = Identifier Id | Number Integer | EAdd Expr Expr | ESub Expr Expr | EShiftL Expr Expr deriving (Read, Show, Eq, Ord) type LocExpr = (Maybe Location, Expr) -- deriving (Read, Show, Eq, Ord) data Instruction = Subleq deriving (Read, Show, Eq, Ord) instructionArity :: Instruction -> (Int, Int) instructionArity Subleq = (1, 3) data Element = ElemInst Instruction [LocExpr] | SubroutineCall (Maybe Location) Id [Expr] | ElemLoc Location deriving (Read, Show, Eq, Ord) data Object = Subroutine Id [Id] [Element] | Macro Id [Id] [Element] deriving (Read, Show, Eq, Ord) data Module = Module (Map Id Object) deriving (Read, Show, Eq, Ord) maybeToSet :: Maybe a -> Set a maybeToSet = maybe S.empty S.singleton elemsSet :: (Ord a)=>Map k a -> Set a elemsSet m = S.fromList (M.elems m) unionsMap :: (Ord b)=>(a -> Set b) -> Map k a -> Set b unionsMap f m = fst $ M.mapAccum (\a b-> (S.union a (f b), ())) S.empty m objectId :: Object -> Id objectId (Subroutine n _ _) = n objectId (Macro n _ _) = n objectArity :: Object -> Int objectArity (Subroutine _ args _) = length args objectArity (Macro _ args _) = length args evaluateNumExpr :: Expr -> Integer evaluateNumExpr (Identifier x) = error $ "unexpected identifier " ++ x ++ "." evaluateNumExpr (Number n) = n evaluateNumExpr (EAdd e1 e2) = evaluateNumExpr e1 + evaluateNumExpr e2 evaluateNumExpr (ESub e1 e2) = evaluateNumExpr e1 - evaluateNumExpr e2 evaluateNumExpr (EShiftL e1 e2) = evaluateNumExpr e1 `shift` fromIntegral (evaluateNumExpr e2) evaluateNumExprInLocElem :: LocExpr -> LocExpr evaluateNumExprInLocElem (l, e) = (l, Number $ evaluateNumExpr e) evaluateNumExprInElem :: Element -> Element evaluateNumExprInElem (ElemInst i les) = ElemInst i $ map evaluateNumExprInLocElem les evaluateNumExprInElem e@(SubroutineCall {}) = e evaluateNumExprInElem e@(ElemLoc {}) = e substituteExpr :: Substitution -> Expr -> Expr substituteExpr sub i@(Identifier x) = M.findWithDefault i x sub substituteExpr sub i@(EAdd e1 e2) = EAdd (substituteExpr sub e1) (substituteExpr sub e2) substituteExpr sub i@(ESub e1 e2) = ESub (substituteExpr sub e1) (substituteExpr sub e2) substituteExpr sub i@(EShiftL e1 e2) = EShiftL (substituteExpr sub e1) (substituteExpr sub e2) substituteExpr _ (Number n) = Number n substituteLocId :: Substitution -> Id -> Id substituteLocId sub l | l `M.member` sub = case M.lookup l sub of Just (Identifier l') -> l' -- Just x -> error $ printf "Label %s cannot be substituted with %s" l (show x) Just x -> printf "%s(%s)" l (show x) -- Just _ -> (Nothing, substituteExpr sub e') Nothing -> l substituteLocId _ l = l substituteLocExpr :: Substitution -> LocExpr -> LocExpr substituteLocExpr sub (Just l, e') = (Just (substituteLocId sub l), substituteExpr sub e') substituteLocExpr sub (l, e') = (l, substituteExpr sub e') substituteElement :: Substitution -> Element -> Element substituteElement sub (ElemInst i es) = ElemInst i (map (substituteLocExpr sub) es) substituteElement sub (SubroutineCall l i es) = SubroutineCall (fmap (substituteLocId sub) l) i (map (substituteExpr sub) es) substituteElement sub (ElemLoc l) = ElemLoc $ substituteLocId sub l substituteObject :: Substitution -> Object -> Object substituteObject sub (Subroutine n args elems) = Subroutine n args $ map (substituteElement sub) elems substituteObject sub (Macro n args elems) = Macro n args $ map (substituteElement sub) elems locationsElement :: Element -> Set Id locationsElement (ElemInst _ es) = S.fromList $ mapMaybe fst es locationsElement (ElemLoc l) = S.singleton l locationsElement (SubroutineCall l _ _) = maybeToSet l locationsElements :: [Element] -> Set Id locationsElements = S.unions . map locationsElement locationsObject :: Object -> Set Id locationsObject (Subroutine _ _ elems) = S.unions $ map locationsElement elems locationsObject (Macro _ _ elems) = S.unions $ map locationsElement elems freqMap :: (Ord a)=>[a] -> M.Map a Int freqMap xs = M.fromListWith (+) . zip xs $ repeat 1 locationsOccursionElement :: Element -> Map Id Int locationsOccursionElement (ElemInst _ es) = freqMap $ mapMaybe fst es locationsOccursionElement (ElemLoc l) = M.singleton l 1 locationsOccursionElement (SubroutineCall Nothing _ _) = M.empty locationsOccursionElement (SubroutineCall (Just l) _ _) = M.singleton l 1 locationsOccursionElements :: [Element] -> Map Id Int locationsOccursionElements = M.unionsWith (+) . map locationsOccursionElement locationsOccursionObject :: Object -> Map Id Int locationsOccursionObject (Subroutine _ _ elems) = locationsOccursionElements elems locationsOccursionObject (Macro _ _ elems) = locationsOccursionElements elems errorsObject :: Object -> [String] errorsObject (Subroutine n args elems) = errorsObject' n args elems errorsObject (Macro n args elems) = errorsObject' n args elems errorsObject' :: Id -> [Id] -> [Element] -> [String] errorsObject' n args elems = catMaybes [e1, e2, e3] where e1 | not . null $ dupLocs = Just $ printf "Object %s: locations must be exclusive, but: %s" n (show dupLocs) | otherwise = Nothing e2 | not . null $ dupArgs = Just $ printf "Object %s: arguments must be exclusive, but: %s" n (show dupArgs) | otherwise = Nothing e3 | not . null $ dupArgLocs = Just $ printf "Object %s: locations and arguments must be exclusive, but: %s" n (show dupArgLocs) | otherwise = Nothing argFreq = freqMap args locFreq = locationsOccursionElements elems dupArgs = M.elems . M.filter (> 1) $ argFreq dupLocs = M.elems . M.filter (> 1) $ locFreq dupArgLocs = M.elems . M.filter (> 1) $ M.unionWith (+) argFreq locFreq boundedVars :: Object -> Set Id boundedVars o@(Subroutine _ args _) = S.fromList args `S.union` locationsObject o boundedVars o@(Macro _ args _) = S.fromList args `S.union` locationsObject o freeVarExpr :: Expr -> Set Id freeVarExpr (Identifier i) = S.singleton i freeVarExpr _ = S.empty freeVarLocExpr :: LocExpr -> Set Id freeVarLocExpr (_,e) = freeVarExpr e freeVarElement :: Element -> Set Id freeVarElement (ElemInst _ es) = S.unions $ map freeVarLocExpr es freeVarElement (SubroutineCall l x es) = S.unions $ [maybeToSet l, S.singleton x] ++ map freeVarExpr es freeVarElement (ElemLoc _) = S.empty freeVarObject :: Object -> Set Id freeVarObject o@(Subroutine _ args es) = S.unions (map freeVarElement es) S.\\ S.fromList args S.\\ locationsObject o freeVarObject o@(Macro _ args es) = S.unions (map freeVarElement es) S.\\ S.fromList args S.\\ locationsObject o freeVarModule :: Module -> Set Id freeVarModule (Module m) = unionsMap freeVarObject m S.\\ M.keysSet m applyObject :: LabelPrefix -> Object -> [Expr] -> [Element] applyObject lp (Macro x as es) = applyObject' lp x as es applyObject _ (Subroutine x _ _) = error $ printf "%s is a subroutine and not applicable" x applyObject' :: LabelPrefix -> Id -> [Id] -> [Element] -> [Expr] -> [Element] applyObject' lp x as es aes | length as == length aes = map (substituteElement sub) $ addLocationPrefix lp targets es -- addLocationPrefix lp $ map (substituteElement sub) es | otherwise = error $ printf "%s takes %d argument(s), but got: %s" x (length as) (show aes) where sub = M.fromList $ zip (map (labelPrefixToString lp ++) as) aes targets = S.fromList as `S.union` locationsElements es type DistinctStack a = ([a], Set a) push :: (Ord a)=>a -> DistinctStack a -> Maybe (DistinctStack a) push x (xs, st) | x `S.member` st = Nothing | otherwise = Just (x:xs, S.insert x st) pop :: (Ord a)=>DistinctStack a -> Maybe (a, DistinctStack a) pop ([], _) = Nothing pop (x:xs, st) = Just (x, (xs, S.delete x st)) emptyStack :: DistinctStack a emptyStack = ([], S.empty) singletonStack :: (Ord a)=>a -> DistinctStack a singletonStack x = ([x], S.singleton x) stackToList :: DistinctStack a -> [a] stackToList = fst lookupModule :: Id -> Module -> Maybe Object lookupModule x (Module m) = M.lookup x m expandMacroAll :: Module -> Module expandMacroAll m@(Module m') = Module $ M.map (expandMacro m) m' expandMacro :: Module -> Object -> Object expandMacro _ o@(Macro {}) = o expandMacro m (Subroutine x as es) = Subroutine x as (concatMap (\(i, e)->expandMacro' (singletonStack x) m [i] e) $ zip [0..] es) expandMacro' :: DistinctStack Id -> Module -> LabelPrefix -> Element -> [Element] expandMacro' stk m lp (SubroutineCall l x as) = es'' where stk' = fromMaybe (error $ printf "%s: Cyclic macro expansion: %s" x (show $ stackToList stk)) (push x stk) o :: Object o = fromMaybe (error $ printf "Object %s is not found in the module: %s" x (show $ stackToList stk)) (lookupModule x m) es' :: [Element] es' = map ElemLoc (maybeToList l) ++ applyObject lp o as es'' = concatMap (\(i, e)-> expandMacro' stk' m (i:lp) e) $ zip [0..] es' expandMacro' _ _ _ e@(ElemInst _ _) = [e] expandMacro' _ _ _ e@(ElemLoc _) = [e] addLocationPrefix :: LabelPrefix -> Set Id -> [Element] -> [Element] addLocationPrefix lp targets elems = elems' where elems' = map (substituteElement sub) elems sub = M.fromSet (Identifier . (labelPrefixToString lp ++)) targets type LabelPrefix = [Int] labelPrefixToString :: LabelPrefix -> String labelPrefixToString = ('_':) . intercalate "_" . reverse . map show
Hara-Laboratory/subleq-toolchain
Language/Subleq/Assembly/Prim.hs
gpl-3.0
10,113
0
13
2,242
3,814
1,957
1,857
187
3
{-# LANGUAGE DeriveGeneric #-} {-# LANGUAGE MultiParamTypeClasses #-} module Ampersand.Output.ToJSON.Relations (Relationz) where import Ampersand.ADL1 import Ampersand.FSpec.FSpecAux import Ampersand.Output.ToJSON.JSONutils import Data.Maybe import qualified Data.Set as Set data Relationz = Relationz [RelationJson]deriving (Generic, Show) data RelationJson = RelationJson { relJSONname :: String , relJSONsignature :: String , relJSONsrcConceptId :: String , relJSONtgtConceptId :: String , relJSONuni :: Bool , relJSONtot :: Bool , relJSONinj :: Bool , relJSONsur :: Bool , relJSONprop :: Bool , relJSONaffectedConjuncts :: [String] , relJSONmysqlTable :: RelTableInfo } deriving (Generic, Show) data RelTableInfo = RelTableInfo -- Contains info about where the relation is implemented in SQL { rtiJSONname :: String , rtiJSONtableOf :: Maybe String -- specifies if relation is administrated in table of srcConcept (i.e. "src"), tgtConcept (i.e. "tgt") or its own n-n table (i.e. null). , rtiJSONsrcCol :: TableCol , rtiJSONtgtCol :: TableCol } deriving (Generic, Show) data TableCol = TableCol { tcJSONname :: String , tcJSONnull :: Bool , tcJSONunique :: Bool } deriving (Generic, Show) instance ToJSON Relationz where toJSON = amp2Jason instance ToJSON RelationJson where toJSON = amp2Jason instance ToJSON RelTableInfo where toJSON = amp2Jason instance ToJSON TableCol where toJSON = amp2Jason instance JSON MultiFSpecs Relationz where fromAmpersand multi _ = Relationz (map (fromAmpersand multi) (Set.elems $ vrels (userFSpec multi))) instance JSON Relation RelationJson where fromAmpersand multi dcl = RelationJson { relJSONname = name dcl , relJSONsignature = name dcl ++ (show . sign) dcl , relJSONsrcConceptId = escapeIdentifier . name . source $ dcl , relJSONtgtConceptId = escapeIdentifier . name . target $ dcl , relJSONuni = isUni bindedExp , relJSONtot = isTot bindedExp , relJSONinj = isInj bindedExp , relJSONsur = isSur bindedExp , relJSONprop = isProp bindedExp , relJSONaffectedConjuncts = map rc_id $ fromMaybe [] (lookup dcl $ allConjsPerDecl fSpec) , relJSONmysqlTable = fromAmpersand multi dcl } where bindedExp = EDcD dcl fSpec = userFSpec multi instance JSON Relation RelTableInfo where fromAmpersand multi dcl = RelTableInfo { rtiJSONname = name plug , rtiJSONtableOf = srcOrtgt , rtiJSONsrcCol = fromAmpersand multi . rsSrcAtt $ relstore , rtiJSONtgtCol = fromAmpersand multi . rsTrgAtt $ relstore } where fSpec = userFSpec multi (plug,relstore) = getRelationTableInfo fSpec dcl (plugSrc,_) = getConceptTableInfo fSpec (source dcl) (plugTrg,_) = getConceptTableInfo fSpec (target dcl) srcOrtgt | (plug == plugSrc) && (plugSrc == plugTrg) = Just $ if rsStoredFlipped relstore then "tgt" else "src" -- relations where src and tgt concepts are in the same classification tree as well as relations that are UNI or INJ | plug == plugSrc = Just "src" -- relation in same table as src concept (UNI relations) | plug == plugTrg = Just "tgt" -- relation in same table as tgt concept (INJ relations that are not UNI) | otherwise = Nothing -- relations in n-n table (not UNI and not INJ) instance JSON SqlAttribute TableCol where fromAmpersand _ att = TableCol { tcJSONname = attName att , tcJSONnull = attDBNull att , tcJSONunique = attUniq att }
AmpersandTarski/ampersand
src/Ampersand/Output/ToJSON/Relations.hs
gpl-3.0
3,737
0
13
956
831
461
370
79
0
import Development.Hake import Development.Hake.FunSetRaw main = hake [ file [ "program" ] [ "main.o", "iodat.o", "dorun.o", "lo.o", "./usr/fred/lib/crtn.a" ] $ const [ [ "cc", "-o", "program", "main.o", "iodat.o", "dorun.o", "lo.o", "./usr/fred/lib/crtn.a" ] ] , file [ "main.o" ] [ "main.c" ] $ const [ [ "cc", "-c", "main.c" ] ] , file [ "iodat.o" ] [ "iodat.c" ] $ const [ [ "cc", "-c", "iodat.c" ] ] , file [ "dorun.o" ] [ "dorun.c" ] $ const [ [ "cc", "-c", "dorun.c" ] ] , file [ "lo.o" ] [ "lo.s" ] $ const [ [ "cc", "-c", "lo.s" ] ] , task "clean" [ [ "rm", "main.o", "iodat.o", "dorun.o", "lo.o", "program" ] ] ]
YoshikuniJujo/hake_haskell
examples/nutshell/chap1/1/hakeMainRaw.hs
gpl-3.0
641
1
9
133
244
142
102
10
1
module Fun.Parser.Pragma where -- Imports de parsec. import Text.Parsec import Text.Parsec.Token( colon, symbol, reservedOp ) import Data.Text ( Text, unpack ) -- Imports de fun. import Fun.Parser.Internal -- Imports de equ. import Equ.Syntax ( Operator(..) ) import qualified Equ.Theories.FOL as F ( theoryOperatorsList ) import qualified Equ.Theories.Arith as A ( theoryOperatorsList ) {- Ejemplo {# AC: ≡ ∧ ∨ #} -} kwPragma :: ParserD () kwPragma = keyword "AC" pColon :: ParserD () pColon = colon lexer >> return () pragmaBetween :: ParserD a -> ParserD a pragmaBetween = between (symbol lexer "{#") (symbol lexer "#}") parsePragma :: ParserD [Operator] parsePragma = try $ pragmaBetween pPragma pPragma :: ParserD [Operator] pPragma = kwPragma >> pColon >> parseOperators operatorsList :: [Operator] operatorsList = F.theoryOperatorsList ++ A.theoryOperatorsList parseOperators :: ParserD [Operator] parseOperators = many1 $ choice $ map makeOp operatorsList makeOp :: Operator -> ParserD Operator makeOp op = try (parseOp $ opRepr op) >> return op where parseOp :: Text -> ParserD () parseOp = reservedOp lexer . unpack
alexgadea/fun
Fun/Parser/Pragma.hs
gpl-3.0
1,169
0
9
205
345
191
154
26
1
<?xml version='1.0' encoding='ISO-8859-1' ?> <!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd"> <helpset version="2.0"> <title>Xena Help</title> <maps> <homeID>xenahelp</homeID> <mapref location="map.jhm"/> </maps> <view xml:lang="en" mergetype="javax.help.UniteAppendMerge"> <name>TOC</name> <label>Table Of Contents</label> <type>javax.help.TOCView</type> <data>toc.xml</data> <image>contents</image> </view> <presentation default=true> <name>Xena Help</name> <size width="800" height="600" /> <location x="200" y="200" /> <title>Xena Help</title> <toolbar> <helpaction image="back">javax.help.BackAction</helpaction> <helpaction image="forward">javax.help.ForwardAction</helpaction> <helpaction image="home">javax.help.HomeAction</helpaction> <helpaction image="print">javax.help.PrintAction</helpaction> <helpaction image="printsetup">javax.help.PrintSetupAction</helpaction> </toolbar> <image>titlebaricon</image> </presentation> </helpset>
srnsw/xena
xena/doc/xena-help.hs
gpl-3.0
1,151
116
13
173
520
237
283
-1
-1
module Sound.Tidal.MIDI.Tanzbar where import Sound.Tidal.Params import Sound.Tidal.MIDI.Control (bd1_attack, bd1_attack_p) = pF "bd1_attack" (Just 0) (bd1_decay, bd1_decay_p) = pF "bd1_decay" (Just 0) (bd1_tune, bd1_tune_p) = pF "bd1_tune" (Just 0) (bd1_noise, bd1_noise_p) = pF "bd1_noise" (Just 0) (bd1_filter, bd1_filter_p) = pF "bd1_filter" (Just 0) (bd1_dist, bd1_dist_p) = pF "bd1_dist" (Just 0) (bd1_trigger, bd1_trigger_p) = pF "bd1_trigger" (Just 0) (bd2_decay, bd2_decay_p) = pF "bd2_decay" (Just 0) (bd2_tune, bd2_tune_p) = pF "bd2_tune" (Just 0) (bd2_tone, bd2_tone_p) = pF "bd2_tone" (Just 0) (sd_tune, sd_tune_p) = pF "sd_tune" (Just 0) (sd_d_tune, sd_d_tune_p) = pF "sd_d_tune" (Just 0) (sd_snappy, sd_snappy_p) = pF "sd_snappy" (Just 0) (sd_sn_decay, sd_sn_decay_p) = pF "sd_sn_decay" (Just 0) (sd_tone, sd_tone_p) = pF "sd_tone" (Just 0) (sd_tone_decay, sd_tone_decay_p) = pF "sd_tone_decay" (Just 0) (sd_pitch, sd_pitch_p) = pF "sd_pitch" (Just 0) (rs_tune, rs_tune_p) = pF "rs_tune" (Just 0) (cy_decay, cy_decay_p) = pF "cy_decay" (Just 0) (cy_tone, cy_tone_p) = pF "cy_tone" (Just 0) (cy_tune, cy_tune_p) = pF "cy_tune" (Just 0) (oh_decay, oh_decay_p) = pF "oh_decay" (Just 0) (hh_tune, hh_tune_p) = pF "hh_tune" (Just 0) (hh_decay, hh_decay_p) = pF "hh_decay" (Just 0) (cl_tune, cl_tune_p) = pF "cl_tune" (Just 0) (cl_decay, cl_decay_p) = pF "cl_decay" (Just 0) (cp_decay, cp_decay_p) = pF "cp_decay" (Just 0) (cp_filter, cp_filter_p) = pF "cp_filter" (Just 0) (cp_attack, cp_attack_p) = pF "cp_attack" (Just 0) (cp_trigger, cp_trigger_p) = pF "cp_trigger" (Just 0) (htc_tune, htc_tune_p) = pF "htc_tune" (Just 0) (htc_decay, htc_decay_p) = pF "htc_decay" (Just 0) (htc_noise_on_off, htc_noise_on_off_p) = pF "htc_noise_on_off" (Just 0) (htc_tom_conga, htc_tom_conga_p) = pF "htc_tom_conga" (Just 0) (mtc_tune, mtc_tune_p) = pF "mtc_tune" (Just 0) (mtc_decay, mtc_decay_p) = pF "mtc_decay" (Just 0) (mtc_noise_on_off, mtc_noise_on_off_p) = pF "mtc_noise_on_off" (Just 0) (mtc_tom_conga, mtc_tom_conga_p) = pF "mtc_tom_conga" (Just 0) (ltc_tune, ltc_tune_p) = pF "ltc_tune" (Just 0) (ltc_decay, ltc_decay_p) = pF "ltc_decay" (Just 0) (ltc_noise_on_off, ltc_noise_on_off_p) = pF "ltc_noise_on_off" (Just 0) (ltc_tom_conga, ltc_tom_conga_p) = pF "ltc_tom_conga" (Just 0) (tom_noise, tom_noise_p) = pF "tom_noise" (Just 0) (cb_tune, cb_tune_p) = pF "cb_tune" (Just 0) (cb_decay, cb_decay_p) = pF "cb_decay" (Just 0) (ma_decay, ma_decay_p) = pF "ma_decay" (Just 0) (set_select, set_select_p) = pF "set_select" (Just 0) (track_delay_cv1, track_delay_cv1_p) = pF "track_delay_cv1" (Just 0) (track_delay_cv23, track_delay_cv23_p) = pF "track_delay_cv23" (Just 0) (track_delay_bd1, track_delay_bd1_p) = pF "track_delay_bd1" (Just 0) (track_delay_bd2, track_delay_bd2_p) = pF "track_delay_bd2" (Just 0) (track_delay_sd, track_delay_sd_p) = pF "track_delay_sd" (Just 0) (track_delay_rs, track_delay_rs_p) = pF "track_delay_rs" (Just 0) (track_delay_cy, track_delay_cy_p) = pF "track_delay_cy" (Just 0) (track_delay_oh, track_delay_oh_p) = pF "track_delay_oh" (Just 0) (track_delay_hh, track_delay_hh_p) = pF "track_delay_hh" (Just 0) (track_delay_cl, track_delay_cl_p) = pF "track_delay_cl" (Just 0) (track_delay_cp, track_delay_cp_p) = pF "track_delay_cp" (Just 0) (track_delay_ltc, track_delay_ltc_p) = pF "track_delay_ltc" (Just 0) (track_delay_mtc, track_delay_mtc_p) = pF "track_delay_mtc" (Just 0) (track_delay_htc, track_delay_htc_p) = pF "track_delay_htc" (Just 0) (track_delay_cb, track_delay_cb_p) = pF "track_delay_cb" (Just 0) (track_delay_ma, track_delay_ma_p) = pF "track_delay_ma" (Just 0) (bd1, bd1_p) = pF "bd1" (Just 0) (bd2, bd2_p) = pF "bd2" (Just 0) (sd, sd_p) = pF "sd" (Just 0) (rs, rs_p) = pF "rs" (Just 0) (cy, cy_p) = pF "cy" (Just 0) (oh, oh_p) = pF "oh" (Just 0) (hh, hh_p) = pF "hh" (Just 0) (cl, cl_p) = pF "cl" (Just 0) (cp, cp_p) = pF "cp" (Just 0) (ltc, ltc_p) = pF "ltc" (Just 0) (mtc, mtc_p) = pF "mtc" (Just 0) (htc, htc_p) = pF "htc" (Just 0) (cb, cb_p) = pF "cb" (Just 0) (ma, ma_p) = pF "ma" (Just 0) tanzController :: ControllerShape tanzController = ControllerShape { controls = [ mCC bd1_attack_p 2, mCC bd1_decay_p 64, mCC bd1_tune_p 3, mCC bd1_noise_p 4, mCC bd1_filter_p 5, mCC bd1_dist_p 6, mCC bd1_trigger_p 66, mCC bd2_decay_p 8, mCC bd2_tune_p 9, mCC bd2_tone_p 10, mCC sd_tune_p 11, mCC sd_d_tune_p 12, mCC sd_snappy_p 13, mCC sd_sn_decay_p 67, mCC sd_tone_p 14, mCC sd_tone_decay_p 68, mCC sd_pitch_p 69, mCC rs_tune_p 88, mCC cy_decay_p 70, mCC cy_tone_p 15, mCC cy_tune_p 71, mCC oh_decay_p 75, mCC hh_tune_p 73, mCC hh_decay_p 74, mCC cl_tune_p 16, mCC cl_decay_p 17, mCC cp_decay_p 75, mCC cp_filter_p 18, mCC cp_attack_p 76, mCC cp_trigger_p 77, mCC htc_tune_p 19, mCC htc_decay_p 20, mCC htc_noise_on_off_p 78, mCC htc_tom_conga_p 79, mCC mtc_tune_p 21, mCC mtc_decay_p 22, mCC mtc_noise_on_off_p 80, mCC mtc_tom_conga_p 81, mCC ltc_tune_p 23, mCC ltc_decay_p 24, mCC ltc_noise_on_off_p 82, mCC ltc_tom_conga_p 83, mCC tom_noise_p 84, mCC cb_tune_p 85, mCC cb_decay_p 86, mCC ma_decay_p 87, mCC set_select_p 0, mCC track_delay_cv1_p 89, mCC track_delay_cv23_p 90, mCC track_delay_bd1_p 91, mCC track_delay_bd2_p 92, mCC track_delay_sd_p 93, mCC track_delay_rs_p 94, mCC track_delay_cy_p 95, mCC track_delay_oh_p 96, mCC track_delay_hh_p 97, mCC track_delay_cl_p 98, mCC track_delay_cp_p 99, mCC track_delay_ltc_p 100, mCC track_delay_mtc_p 101, mCC track_delay_htc_p 102, mCC track_delay_cb_p 103, mCC track_delay_ma_p 104, mCC bd1_p 36, mCC bd2_p 37, mCC sd_p 38, mCC rs_p 39, mCC cy_p 40, mCC oh_p 41, mCC hh_p 42, mCC cl_p 43, mCC cp_p 44, mCC ltc_p 45, mCC mtc_p 46, mCC htc_p 47, mCC cb_p 48, mCC ma_p 49 ], latency = 0.1 } tanz = midinote . (tanzN <$>) tanzN :: String -> Int tanzN "bd1" = 36 tanzN "bd2" = 37 tanzN "sd" = 38 tanzN "rs" = 39 tanzN "cy" = 40 tanzN "oh" = 41 tanzN "hh" = 42 tanzN "cl" = 43 tanzN "cp" = 44 tanzN "ltc" = 45 tanzN "mtc" = 46 tanzN "htc" = 47 tanzN "cb" = 48 tanzN "ma" = 49 -- general shape for stream tanzShape = toShape tanzController
tidalcycles/tidal-midi
Sound/Tidal/MIDI/Tanzbar.hs
gpl-3.0
6,380
0
8
1,207
2,575
1,336
1,239
178
1
module Paths_lambda ( version, getBinDir, getLibDir, getDataDir, getLibexecDir, getDataFileName, getSysconfDir ) where import qualified Control.Exception as Exception import Data.Version (Version(..)) import System.Environment (getEnv) import Prelude catchIO :: IO a -> (Exception.IOException -> IO a) -> IO a catchIO = Exception.catch version :: Version version = Version {versionBranch = [0,1,0,0], versionTags = []} bindir, libdir, datadir, libexecdir, sysconfdir :: FilePath bindir = "/home/anastasia/Haskell/lambda/.cabal-sandbox/bin" libdir = "/home/anastasia/Haskell/lambda/.cabal-sandbox/lib/x86_64-linux-ghc-7.8.3/lambda-0.1.0.0" datadir = "/home/anastasia/Haskell/lambda/.cabal-sandbox/share/x86_64-linux-ghc-7.8.3/lambda-0.1.0.0" libexecdir = "/home/anastasia/Haskell/lambda/.cabal-sandbox/libexec" sysconfdir = "/home/anastasia/Haskell/lambda/.cabal-sandbox/etc" getBinDir, getLibDir, getDataDir, getLibexecDir, getSysconfDir :: IO FilePath getBinDir = catchIO (getEnv "lambda_bindir") (\_ -> return bindir) getLibDir = catchIO (getEnv "lambda_libdir") (\_ -> return libdir) getDataDir = catchIO (getEnv "lambda_datadir") (\_ -> return datadir) getLibexecDir = catchIO (getEnv "lambda_libexecdir") (\_ -> return libexecdir) getSysconfDir = catchIO (getEnv "lambda_sysconfdir") (\_ -> return sysconfdir) getDataFileName :: FilePath -> IO FilePath getDataFileName name = do dir <- getDataDir return (dir ++ "/" ++ name)
nastya13/lambda
dist/build/autogen/Paths_lambda.hs
gpl-3.0
1,466
0
10
182
371
213
158
28
1
module He.Error ( module Control.Monad.Except , Error() , err , err' , report , check , checked , fatal , logExceptT , logExceptT' , log ) where import Control.Lens import Control.Monad.Except import qualified Data.DList as D import H.Prelude hiding ((<>), empty, show) import Prelude (show) import Text.Parsec.Applicative.Types import Text.PrettyPrint import Text.PrettyPrint.HughesPJClass import System.IO (hPutStr, hPutStrLn, stderr) newtype Error = Error (D.DList (Maybe SourcePos, Doc)) deriving (Eq, Monoid, Show) vSepBy :: Doc -> Doc -> Doc -> Doc vSepBy sep above below = above $+$ sep $+$ below instance Pretty Error where pPrint (Error es) = case D.toList es of [] -> text "No errors" es -> foldl (vSepBy empty) empty $ map (uncurry prettyError) es prettyError :: Maybe SourcePos -> Doc -> Doc prettyError sp message = case sp of Nothing -> message Just sp -> prettySourcePos sp $+$ message prettySourcePos :: SourcePos -> Doc prettySourcePos sp = text "At " <> maybe empty (text . unpack) (sp ^. spName) <> text " line " <> text (show $ sp ^. spLine) <> text " column " <> text (show $ sp ^. spColumn) <> text ":" err :: Maybe SourcePos -> Text -> Error err sp xs = Error $ D.singleton (sp, text $ unpack xs) err' :: Text -> Error err' = err Nothing report :: (MonadState (Maybe e) m, Monoid e) => e -> m () report e = modify (`mappend` Just e) check :: (MonadState (Maybe e) m, MonadError e m) => m () check = get >>= maybe (return ()) ((put Nothing >>) . throwError) checked :: (MonadState (Maybe e) m, MonadError e m) => m a -> m a checked = (<* check) fatal :: (MonadState (Maybe e) m, Monoid e, MonadError e m) => e -> m a fatal e = do report e get >>= maybe (impossible "fatal") ((put Nothing >>) . throwError) logExceptT :: (MonadIO m, Pretty e) => ExceptT e m a -> (a -> m ()) -> m () logExceptT e f = runExceptT e >>= either (liftIO . hPutStr stderr . render . pPrint) f logExceptT' :: (MonadIO m, Pretty e) => ExceptT e m () -> m () logExceptT' e = logExceptT e return log :: (MonadIO m) => Text -> m () log = liftIO . hPutStrLn stderr . unpack
ktvoelker/helium
src/He/Error.hs
gpl-3.0
2,139
0
13
457
946
496
450
-1
-1
import System.IO import System.FilePath import Data.List import Data.List.Split main :: IO () main = do contents <- readFile "..\\goldi_input_1.txt" let (pref:orgs) = zip [0..] $ conv $ getGoldi contents putStrLn $ intercalate " " $ map (show . fst) $ filter (checkSeats (snd pref)) $ orgs checkSeats :: (Int, Int) -> (Int, (Int, Int)) -> Bool checkSeats a b = ((fst a) <= (fst $ snd b)) && ((snd a) >= (snd $ snd b)) getGoldi :: String -> [ [Char]] getGoldi contents = [ x | y <- lines contents, x <- splitOn " " y] conv :: [[Char]] -> [ (Int, Int) ] conv [] = [] conv (k:v:t) = (read k :: Int, read v :: Int) : conv t
jandersen7/Daily
src/291e/hs/goldi.hs
gpl-3.0
639
0
13
142
370
187
183
16
1
{- ============================================================================ | Copyright 2011 Matthew D. Steele <[email protected]> | | | | This file is part of Fallback. | | | | Fallback is free software: you can redistribute it and/or modify it under | | the terms of the GNU General Public License as published by the Free | | Software Foundation, either version 3 of the License, or (at your option) | | any later version. | | | | Fallback is distributed in the hope that it will be useful, but WITHOUT | | ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | | FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for | | more details. | | | | You should have received a copy of the GNU General Public License along | | with Fallback. If not, see <http://www.gnu.org/licenses/>. | ============================================================================ -} module Fallback.Draw (module Fallback.Draw.Base, module Fallback.Draw.GUI) where import Fallback.Draw.Base import Fallback.Draw.GUI -------------------------------------------------------------------------------
mdsteele/fallback
src/Fallback/Draw.hs
gpl-3.0
1,660
0
5
696
36
25
11
5
0
-- grid is a game written in Haskell -- Copyright (C) 2018 [email protected] -- -- This file is part of grid. -- -- grid is free software: you can redistribute it and/or modify -- it under the terms of the GNU General Public License as published by -- the Free Software Foundation, either version 3 of the License, or -- (at your option) any later version. -- -- grid is distributed in the hope that it will be useful, -- but WITHOUT ANY WARRANTY; without even the implied warranty of -- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -- GNU General Public License for more details. -- -- You should have received a copy of the GNU General Public License -- along with grid. If not, see <http://www.gnu.org/licenses/>. -- {-# LANGUAGE MagicHash #-} {-# LANGUAGE UnboxedTuples #-} module Game.Grid.GridWorld.SegmentArray ( SegmentArray, makeSegmentArray, segmentarrayRead, segmentarrayWrite, ) where import MyPrelude import Game import Game.Grid.GridWorld.Segment import Game.Grid.GridWorld.Node import Game.Grid.GridWorld.Turn import Data.Int -- making unportable import GHC.Prim import GHC.Exts import GHC.Types import GHC.Int import System.IO.Unsafe -------------------------------------------------------------------------------- -- SegmentArray -- we need some type of sequential container for the segments of path. -- also, we want the path to be able to grow modulo its size. so I decided -- to use a custom array type. -- update: after working in IO, we could have used IOArray instead, but read note -- below -- note: in GHC, the datatypes Int8/Int16/... are implemented by the word size -- of the architecture (which is 32 bits on ARMv7), so using them does not -- save memory. but in this implementation of SegmentArray, we use the cor- -- responding bits and so save memory. data SegmentArray = SegmentArray (MutableByteArray# RealWorld) unsafeState# :: (State# RealWorld -> (# State# RealWorld, a #)) -> a unsafeState# = unsafeDupablePerformIO . IO makeSegmentArray :: UInt -> SegmentArray makeSegmentArray (W# size#) = unsafeState# $ \s# -> case newAlignedPinnedByteArray# (word2Int# (size# `timesWord#` 16##)) 4# s# of (# s'#, mba# #) -> (# s'#, SegmentArray mba# #) segmentarrayWrite :: SegmentArray -> UInt -> Segment -> SegmentArray segmentarrayWrite sa@(SegmentArray mba#) (W# ix#) (Segment (Node (I16# x#) (I16# y#) (I16# z#)) (Turn (I8# x0#) (I8# x1#) (I8# x2#) (I8# y0#) (I8# y1#) (I8# y2#) (I8# z0#) (I8# z1#) (I8# z2#))) = unsafeState# $ \s# -> case writeInt16Array# mba# (ix2Bytes# ix# 0##) x# s# of { s# -> case writeInt16Array# mba# (ix2Bytes# ix# 1##) y# s# of { s# -> case writeInt16Array# mba# (ix2Bytes# ix# 2##) z# s# of { s# -> case writeInt8Array# mba# (ix4Bytes# ix# 6##) x0# s# of { s# -> case writeInt8Array# mba# (ix4Bytes# ix# 7##) x1# s# of { s# -> case writeInt8Array# mba# (ix4Bytes# ix# 8##) x2# s# of { s# -> case writeInt8Array# mba# (ix4Bytes# ix# 9##) y0# s# of { s# -> case writeInt8Array# mba# (ix4Bytes# ix# 10##) y1# s# of { s# -> case writeInt8Array# mba# (ix4Bytes# ix# 11##) y2# s# of { s# -> case writeInt8Array# mba# (ix4Bytes# ix# 12##) z0# s# of { s# -> case writeInt8Array# mba# (ix4Bytes# ix# 13##) z1# s# of { s# -> case writeInt8Array# mba# (ix4Bytes# ix# 14##) z2# s# of { s# -> case writeInt8Array# mba# (ix4Bytes# ix# 15##) 0# s# of { s# -> (# s#, sa #) }}}}}}}}}}}}} where ix2Bytes# ix# a# = word2Int# ((8## `timesWord#` ix#) `plusWord#` a#) ix4Bytes# ix# a# = word2Int# ((16## `timesWord#` ix#) `plusWord#` a#) segmentarrayRead :: SegmentArray -> UInt -> Segment segmentarrayRead (SegmentArray mba#) (W# ix#) = unsafeState# $ \s# -> case readInt16Array# mba# (ix2Bytes# ix# 0##) s# of { (# s#, x# #) -> case readInt16Array# mba# (ix2Bytes# ix# 1##) s# of { (# s#, y# #) -> case readInt16Array# mba# (ix2Bytes# ix# 2##) s# of { (# s#, z# #) -> case readInt8Array# mba# (ix4Bytes# ix# 6##) s# of { (# s#, x0# #) -> case readInt8Array# mba# (ix4Bytes# ix# 7##) s# of { (# s#, x1# #) -> case readInt8Array# mba# (ix4Bytes# ix# 8##) s# of { (# s#, x2# #) -> case readInt8Array# mba# (ix4Bytes# ix# 9##) s# of { (# s#, y0# #) -> case readInt8Array# mba# (ix4Bytes# ix# 10##) s# of { (# s#, y1# #) -> case readInt8Array# mba# (ix4Bytes# ix# 11##) s# of { (# s#, y2# #) -> case readInt8Array# mba# (ix4Bytes# ix# 12##) s# of { (# s#, z0# #) -> case readInt8Array# mba# (ix4Bytes# ix# 13##) s# of { (# s#, z1# #) -> case readInt8Array# mba# (ix4Bytes# ix# 14##) s# of { (# s#, z2# #) -> case readInt8Array# mba# (ix4Bytes# ix# 15##) s# of { (# s#, _ #) -> (# s#, Segment (Node (I16# x#) (I16# y#) (I16# z#)) (Turn (I8# x0#) (I8# x1#) (I8# x2#) (I8# y0#) (I8# y1#) (I8# y2#) (I8# z0#) (I8# z1#) (I8# z2#)) #) }}}}}}}}}}}}} where ix2Bytes# ix# a# = word2Int# ((8## `timesWord#` ix#) `plusWord#` a#) ix4Bytes# ix# a# = word2Int# ((16## `timesWord#` ix#) `plusWord#` a#) {- -- | we create a newtype so that we later can just store the direction if we want -- (strip Segment for structure) newtype StorableSegment = StorableSegment Segment wrapStorableSegment :: Segment -> StorableSegment wrapStorableSegment seg = StorableSegment seg instance Storable StorableSegment where sizeOf _ = 16 alignment _ = 4 -- ^ 4 byte align on ARM (?) peek ptr = do -- node x <- peekByteOff ptr 0 :: IO Int16 y <- peekByteOff ptr 2 :: IO Int16 z <- peekByteOff ptr 4 :: IO Int16 -- turn x0 <- peekByteOff ptr 6 :: IO Int8 x1 <- peekByteOff ptr 7 :: IO Int8 x2 <- peekByteOff ptr 8 :: IO Int8 y0 <- peekByteOff ptr 9 :: IO Int8 y1 <- peekByteOff ptr 10 :: IO Int8 y2 <- peekByteOff ptr 11 :: IO Int8 z0 <- peekByteOff ptr 12 :: IO Int8 z1 <- peekByteOff ptr 13 :: IO Int8 z2 <- peekByteOff ptr 14 :: IO Int8 -- 15 is empty! return $ StorableSegment $ Segment (Node x y z) (Turn x0 x1 x2 y0 y1 y2 z0 z1 z2) poke ptr (StorableSegment (Segment (Node x y z) (Turn x0 x1 x2 y0 y1 y2 z0 z1 z2))) = do -- node pokeByteOff ptr 0 x pokeByteOff ptr 2 y pokeByteOff ptr 4 z -- turn pokeByteOff ptr 6 x0 pokeByteOff ptr 7 x1 pokeByteOff ptr 8 x2 pokeByteOff ptr 9 y0 pokeByteOff ptr 10 y1 pokeByteOff ptr 11 y2 pokeByteOff ptr 12 z0 pokeByteOff ptr 13 z1 pokeByteOff ptr 14 z2 -- 15: empty! -}
karamellpelle/grid
source/Game/Grid/GridWorld/SegmentArray.hs
gpl-3.0
7,214
0
51
2,150
1,491
812
679
76
1
{-# LANGUAGE DataKinds #-} {-# LANGUAGE DeriveGeneric #-} {-# LANGUAGE FlexibleInstances #-} {-# LANGUAGE GeneralizedNewtypeDeriving #-} {-# LANGUAGE LambdaCase #-} {-# LANGUAGE NoImplicitPrelude #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE RecordWildCards #-} {-# LANGUAGE TypeFamilies #-} {-# OPTIONS_GHC -fno-warn-unused-imports #-} -- Module : Network.AWS.OpsWorks.DeregisterInstance -- Copyright : (c) 2013-2014 Brendan Hay <[email protected]> -- License : This Source Code Form is subject to the terms of -- the Mozilla Public License, v. 2.0. -- A copy of the MPL can be found in the LICENSE file or -- you can obtain it at http://mozilla.org/MPL/2.0/. -- Maintainer : Brendan Hay <[email protected]> -- Stability : experimental -- Portability : non-portable (GHC extensions) -- -- Derived from AWS service descriptions, licensed under Apache 2.0. -- | Deregister a registered Amazon EC2 or on-premises instance. This action -- removes the instance from the stack and returns it to your control. This -- action can not be used with instances that were created with AWS OpsWorks. -- -- Required Permissions: To use this action, an IAM user must have a Manage -- permissions level for the stack or an attached policy that explicitly grants -- permissions. For more information on user permissions, see <http://docs.aws.amazon.com/opsworks/latest/userguide/opsworks-security-users.html Managing UserPermissions>. -- -- <http://docs.aws.amazon.com/opsworks/latest/APIReference/API_DeregisterInstance.html> module Network.AWS.OpsWorks.DeregisterInstance ( -- * Request DeregisterInstance -- ** Request constructor , deregisterInstance -- ** Request lenses , di1InstanceId -- * Response , DeregisterInstanceResponse -- ** Response constructor , deregisterInstanceResponse ) where import Network.AWS.Prelude import Network.AWS.Request.JSON import Network.AWS.OpsWorks.Types import qualified GHC.Exts newtype DeregisterInstance = DeregisterInstance { _di1InstanceId :: Text } deriving (Eq, Ord, Read, Show, Monoid, IsString) -- | 'DeregisterInstance' constructor. -- -- The fields accessible through corresponding lenses are: -- -- * 'di1InstanceId' @::@ 'Text' -- deregisterInstance :: Text -- ^ 'di1InstanceId' -> DeregisterInstance deregisterInstance p1 = DeregisterInstance { _di1InstanceId = p1 } -- | The instance ID. di1InstanceId :: Lens' DeregisterInstance Text di1InstanceId = lens _di1InstanceId (\s a -> s { _di1InstanceId = a }) data DeregisterInstanceResponse = DeregisterInstanceResponse deriving (Eq, Ord, Read, Show, Generic) -- | 'DeregisterInstanceResponse' constructor. deregisterInstanceResponse :: DeregisterInstanceResponse deregisterInstanceResponse = DeregisterInstanceResponse instance ToPath DeregisterInstance where toPath = const "/" instance ToQuery DeregisterInstance where toQuery = const mempty instance ToHeaders DeregisterInstance instance ToJSON DeregisterInstance where toJSON DeregisterInstance{..} = object [ "InstanceId" .= _di1InstanceId ] instance AWSRequest DeregisterInstance where type Sv DeregisterInstance = OpsWorks type Rs DeregisterInstance = DeregisterInstanceResponse request = post "DeregisterInstance" response = nullResponse DeregisterInstanceResponse
dysinger/amazonka
amazonka-opsworks/gen/Network/AWS/OpsWorks/DeregisterInstance.hs
mpl-2.0
3,538
0
9
707
351
216
135
47
1
{-# LANGUAGE DataKinds #-} {-# LANGUAGE DeriveGeneric #-} {-# LANGUAGE FlexibleInstances #-} {-# LANGUAGE GeneralizedNewtypeDeriving #-} {-# LANGUAGE LambdaCase #-} {-# LANGUAGE NoImplicitPrelude #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE RecordWildCards #-} {-# LANGUAGE TypeFamilies #-} {-# OPTIONS_GHC -fno-warn-unused-imports #-} -- Module : Network.AWS.SWF.DeprecateWorkflowType -- Copyright : (c) 2013-2014 Brendan Hay <[email protected]> -- License : This Source Code Form is subject to the terms of -- the Mozilla Public License, v. 2.0. -- A copy of the MPL can be found in the LICENSE file or -- you can obtain it at http://mozilla.org/MPL/2.0/. -- Maintainer : Brendan Hay <[email protected]> -- Stability : experimental -- Portability : non-portable (GHC extensions) -- -- Derived from AWS service descriptions, licensed under Apache 2.0. -- | Deprecates the specified /workflow type/. After a workflow type has been -- deprecated, you cannot create new executions of that type. Executions that -- were started before the type was deprecated will continue to run. A -- deprecated workflow type may still be used when calling visibility actions. -- -- This operation is eventually consistent. The results are best effort and may -- not exactly reflect recent updates and changes. Access Control -- -- You can use IAM policies to control this action's access to Amazon SWF -- resources as follows: -- -- Use a 'Resource' element with the domain name to limit the action to only -- specified domains. Use an 'Action' element to allow or deny permission to call -- this action. Constrain the following parameters by using a 'Condition' element -- with the appropriate keys. 'workflowType.name': String constraint. The key is 'swf:workflowType.name'. 'workflowType.version': String constraint. The key is 'swf:workflowType.version' -- . If the caller does not have sufficient permissions to invoke the action, -- or the parameter values fall outside the specified constraints, the action -- fails. The associated event attribute's cause parameter will be set to -- OPERATION_NOT_PERMITTED. For details and example IAM policies, see <http://docs.aws.amazon.com/amazonswf/latest/developerguide/swf-dev-iam.html Using IAMto Manage Access to Amazon SWF Workflows>. -- -- <http://docs.aws.amazon.com/amazonswf/latest/apireference/API_DeprecateWorkflowType.html> module Network.AWS.SWF.DeprecateWorkflowType ( -- * Request DeprecateWorkflowType -- ** Request constructor , deprecateWorkflowType -- ** Request lenses , dwt1Domain , dwt1WorkflowType -- * Response , DeprecateWorkflowTypeResponse -- ** Response constructor , deprecateWorkflowTypeResponse ) where import Network.AWS.Prelude import Network.AWS.Request.JSON import Network.AWS.SWF.Types import qualified GHC.Exts data DeprecateWorkflowType = DeprecateWorkflowType { _dwt1Domain :: Text , _dwt1WorkflowType :: WorkflowType } deriving (Eq, Read, Show) -- | 'DeprecateWorkflowType' constructor. -- -- The fields accessible through corresponding lenses are: -- -- * 'dwt1Domain' @::@ 'Text' -- -- * 'dwt1WorkflowType' @::@ 'WorkflowType' -- deprecateWorkflowType :: Text -- ^ 'dwt1Domain' -> WorkflowType -- ^ 'dwt1WorkflowType' -> DeprecateWorkflowType deprecateWorkflowType p1 p2 = DeprecateWorkflowType { _dwt1Domain = p1 , _dwt1WorkflowType = p2 } -- | The name of the domain in which the workflow type is registered. dwt1Domain :: Lens' DeprecateWorkflowType Text dwt1Domain = lens _dwt1Domain (\s a -> s { _dwt1Domain = a }) -- | The workflow type to deprecate. dwt1WorkflowType :: Lens' DeprecateWorkflowType WorkflowType dwt1WorkflowType = lens _dwt1WorkflowType (\s a -> s { _dwt1WorkflowType = a }) data DeprecateWorkflowTypeResponse = DeprecateWorkflowTypeResponse deriving (Eq, Ord, Read, Show, Generic) -- | 'DeprecateWorkflowTypeResponse' constructor. deprecateWorkflowTypeResponse :: DeprecateWorkflowTypeResponse deprecateWorkflowTypeResponse = DeprecateWorkflowTypeResponse instance ToPath DeprecateWorkflowType where toPath = const "/" instance ToQuery DeprecateWorkflowType where toQuery = const mempty instance ToHeaders DeprecateWorkflowType instance ToJSON DeprecateWorkflowType where toJSON DeprecateWorkflowType{..} = object [ "domain" .= _dwt1Domain , "workflowType" .= _dwt1WorkflowType ] instance AWSRequest DeprecateWorkflowType where type Sv DeprecateWorkflowType = SWF type Rs DeprecateWorkflowType = DeprecateWorkflowTypeResponse request = post "DeprecateWorkflowType" response = nullResponse DeprecateWorkflowTypeResponse
dysinger/amazonka
amazonka-swf/gen/Network/AWS/SWF/DeprecateWorkflowType.hs
mpl-2.0
4,915
0
9
962
423
264
159
54
1
{-# LANGUAGE OverloadedStrings #-} -- -- Copyright (c) 2005-2022 Stefan Wehr - http://www.stefanwehr.de -- -- This library is free software; you can redistribute it and/or -- modify it under the terms of the GNU Lesser General Public -- License as published by the Free Software Foundation; either -- version 2.1 of the License, or (at your option) any later version. -- -- This library is distributed in the hope that it will be useful, -- but WITHOUT ANY WARRANTY; without even the implied warranty of -- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -- Lesser General Public License for more details. -- -- You should have received a copy of the GNU Lesser General Public -- License along with this library; if not, write to the Free Software -- Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307, USA -- {-| This module defines functions for notifying all test reporters registered about particular events in the lifecycle of a test run. Further, it defines the standard test reporters for HTF's various output formats. -} module Test.Framework.TestReporter ( IsParallel(..), isParallelFromBool, IsJsonOutput(..), IsXmlOutput(..), reportAllTests, reportGlobalStart, reportTestStart, reportTestResult, reportGlobalResults, defaultTestReporters ) where import Test.Framework.TestTypes import Test.Framework.Location import Test.Framework.Colors import Test.Framework.JsonOutput import Test.Framework.XmlOutput import System.IO import Control.Monad.RWS import Text.PrettyPrint import qualified Data.Text.IO as T import qualified Data.ByteString.Lazy as BSL -- | Invokes 'tr_reportAllTests' on all test reporters registered. reportAllTests :: ReportAllTests reportAllTests tests = do reps <- asks tc_reporters mapM_ (\r -> tr_reportAllTests r tests) reps -- | Invokes 'tr_reportGlobalStart' on all test reporters registered. reportGlobalStart :: ReportGlobalStart reportGlobalStart tests = do reps <- asks tc_reporters mapM_ (\r -> tr_reportGlobalStart r tests) reps -- | Invokes 'tr_reportTestStart' on all test reporters registered. reportTestStart :: ReportTestStart reportTestStart t = do reps <- asks tc_reporters mapM_ (\r -> tr_reportTestStart r t) reps -- | Invokes 'tr_reportTestResult' on all test reporters registered. reportTestResult :: ReportTestResult reportTestResult t = do reps <- asks tc_reporters mapM_ (\r -> tr_reportTestResult r t) reps -- | Invokes 'tr_reportGlobalResults' on all test reporters registered. reportGlobalResults :: ReportGlobalResults reportGlobalResults arg = do reps <- asks tc_reporters mapM_ (\r -> tr_reportGlobalResults r arg) reps data IsParallel = Parallel | NonParallel isParallelFromBool :: Bool -> IsParallel isParallelFromBool True = Parallel isParallelFromBool False = NonParallel data IsJsonOutput = JsonOutput | NoJsonOutput data IsXmlOutput = XmlOutput | NoXmlOutput -- | The default test reporters for HTF. defaultTestReporters :: IsParallel -> IsJsonOutput -> IsXmlOutput -> [TestReporter] defaultTestReporters inParallel forMachine doXml = case (inParallel, forMachine) of (NonParallel, NoJsonOutput) -> [TestReporter { tr_id = "rep_seq_human" , tr_reportAllTests = reportAllTestsH , tr_reportGlobalStart = reportGlobalStartHS , tr_reportTestStart = reportTestStartHS , tr_reportTestResult = reportTestResultHS , tr_reportGlobalResults = reportGlobalResultsH }] ++ xmlReporters (Parallel, NoJsonOutput) -> [TestReporter { tr_id = "rep_par_human" , tr_reportAllTests = reportAllTestsH , tr_reportGlobalStart = reportGlobalStartHP , tr_reportTestStart = reportTestStartHP , tr_reportTestResult = reportTestResultHP , tr_reportGlobalResults = reportGlobalResultsH }] ++ xmlReporters (NonParallel, JsonOutput) -> [TestReporter { tr_id = "rep_seq_machine" , tr_reportAllTests = reportAllTestsM , tr_reportGlobalStart = reportGlobalStartMS , tr_reportTestStart = reportTestStartMS , tr_reportTestResult = reportTestResultMS , tr_reportGlobalResults = reportGlobalResultsM }] ++ xmlReporters (Parallel, JsonOutput) -> [TestReporter { tr_id = "rep_par_machine" , tr_reportAllTests = reportAllTestsM , tr_reportGlobalStart = reportGlobalStartMP , tr_reportTestStart = reportTestStartMP , tr_reportTestResult = reportTestResultMP , tr_reportGlobalResults = reportGlobalResultsM }] ++ xmlReporters where xmlReporters = case doXml of NoXmlOutput -> [] XmlOutput -> [(emptyTestReporter "rep_xml") { tr_reportGlobalResults = reportGlobalResultsXml }] -- -- output for humans -- humanTestName :: GenFlatTest a -> String humanTestName ft = flatName (ft_path ft) ++ case ft_location ft of Nothing -> "" Just loc -> " (" ++ showLoc loc ++ ")" reportHumanTestStartMessage :: ReportLevel -> GenFlatTest a -> TR () reportHumanTestStartMessage level ft = do let t = colorize testStartColor "[TEST] " reportTR level (t +++ noColor (humanTestName ft)) -- sequential reportGlobalStartHS :: ReportGlobalStart reportGlobalStartHS _ = return () reportTestStartHS :: ReportTestStart reportTestStartHS ft = reportHumanTestStartMessage Debug ft reportTestResultHS :: ReportTestResult reportTestResultHS ftr = let res = rr_result (ft_payload ftr) msg = attachCallStack (rr_message (ft_payload ftr)) (rr_stack (ft_payload ftr)) in case res of Pass -> reportMessage Debug msg okSuffix Pending -> do reportHumanTestStartMessageIfNeeded reportMessage Info msg pendingSuffix Fail -> do reportHumanTestStartMessageIfNeeded reportMessage Info msg failureSuffix Error -> do reportHumanTestStartMessageIfNeeded reportMessage Info msg errorSuffix where reportHumanTestStartMessageIfNeeded = do tc <- ask when (tc_quiet tc) (reportHumanTestStartMessage Info ftr) reportMessage level msg suffix = reportTR level (ensureNewlineColorString msg +++ suffix +++ noColor timeStr) timeStr = " (" ++ show (rr_wallTimeMs (ft_payload ftr)) ++ "ms)\n" failureSuffix = colorize warningColor "*** Failed!" errorSuffix = colorize warningColor "@@@ Error!" pendingSuffix = colorize pendingColor "^^^ Pending!" okSuffix = colorize testOkColor "+++ OK" -- parallel reportGlobalStartHP :: ReportGlobalStart reportGlobalStartHP _ = return () reportTestStartHP :: ReportTestStart reportTestStartHP ft = do reportStringTR Debug ("Starting " ++ (humanTestName ft)) reportTestResultHP :: ReportTestResult reportTestResultHP ftr = do reportHumanTestStartMessage Debug ftr reportTestResultHS ftr -- results and all tests reportAllTestsH :: ReportAllTests reportAllTestsH l = reportStringTR Info (render (renderTestNames l)) reportGlobalResultsH :: ReportGlobalResults reportGlobalResultsH arg = do let passed = length (rgra_passed arg) pending = length (rgra_pending arg) failed = length (rgra_failed arg) error = length (rgra_errors arg) timedOut = length (rgra_timedOut arg) filtered = length (rgra_filtered arg) total = passed + failed + error + pending let pendings = (if pending > 0 then colorize pendingColor else noColor) "* Pending:" failures = (if failed > 0 then colorize warningColor else noColor) "* Failures:" errors = (if error > 0 then colorize warningColor else noColor) "* Errors:" reportTR Info ("* Tests: " +++ showC total +++ "\n" +++ "* Passed: " +++ showC passed +++ "\n" +++ pendings +++ " " +++ showC pending +++ "\n" +++ failures +++ " " +++ showC failed +++ "\n" +++ errors +++ " " +++ showC error +++ "\n" +++ "* Timed out: " +++ showC timedOut +++ "\n" +++ "* Filtered: " +++ showC filtered) when (timedOut > 0) $ if timedOut < 10 then reportTR Info ("\n" +++ noColor "* Timed out:" +++ "\n" +++ renderTestNames' (reverse (rgra_timedOut arg))) else reportTR Info ("\n" +++ noColor "* Timed out: (" +++ showC timedOut +++ noColor ", too many to list)") when (filtered > 0) $ if filtered < 10 then reportTR Info ("\n" +++ noColor "* Filtered:" +++ "\n" +++ renderTestNames' (reverse (rgra_filtered arg))) else reportTR Info ("\n" +++ noColor "* Filtered: (" +++ showC filtered +++ noColor ", too many to list)") when (pending > 0) $ reportTR Info ("\n" +++ pendings +++ "\n" +++ renderTestNames' (reverse (rgra_pending arg))) when (failed > 0) $ reportTR Info ("\n" +++ failures +++ "\n" +++ renderTestNames' (reverse (rgra_failed arg))) when (error > 0) $ reportTR Info ("\n" +++ errors +++ "\n" +++ renderTestNames' (reverse (rgra_errors arg))) reportStringTR Info ("\nTotal execution time: " ++ show (rgra_timeMs arg) ++ "ms") where showC x = noColor (show x) renderTestNames' rrs = noColor $ render $ nest 2 $ renderTestNames rrs renderTestNames :: [GenFlatTest a] -> Doc renderTestNames l = vcat (map (\ft -> text "*" <+> text (humanTestName ft)) l) -- -- output for machines -- -- sequential reportGlobalStartMS :: ReportGlobalStart reportGlobalStartMS _ = return () reportTestStartMS :: ReportTestStart reportTestStartMS ft = let json = mkTestStartEventObj ft (flatName (ft_path ft)) in reportJsonTR json reportTestResultMS :: ReportTestResult reportTestResultMS ftr = let json = mkTestEndEventObj ftr (flatName (ft_path ftr)) in reportJsonTR json -- parallel reportGlobalStartMP :: ReportGlobalStart reportGlobalStartMP _ = return () reportTestStartMP :: ReportTestStart reportTestStartMP = reportTestStartMS reportTestResultMP :: ReportTestResult reportTestResultMP = reportTestResultMS -- results and all tests reportAllTestsM :: ReportAllTests reportAllTestsM l = let json = mkTestListObj (map (\ft -> (ft, flatName (ft_path ft))) l) in reportJsonTR json reportGlobalResultsM :: ReportGlobalResults reportGlobalResultsM arg = let json = mkTestResultsObj arg in reportJsonTR json reportGlobalResultsXml :: ReportGlobalResults reportGlobalResultsXml arg = do let xml = mkGlobalResultsXml arg tc <- ask case tc_outputXml tc of Just fname -> liftIO $ withFile fname WriteMode $ \h -> BSL.hPut h xml Nothing -> liftIO $ BSL.putStr xml -- -- General reporting routines -- reportTR :: ReportLevel -> ColorString -> TR () reportTR level msg = do tc <- ask let s = renderColorString msg (tc_useColors tc) reportGen tc level (\h -> T.hPutStrLn h s) reportStringTR :: ReportLevel -> String -> TR () reportStringTR level msg = do tc <- ask reportGen tc level (\h -> hPutStrLn h msg) reportLazyBytesTR :: ReportLevel -> BSL.ByteString -> TR () reportLazyBytesTR level msg = do tc <- ask reportGen tc level (\h -> BSL.hPut h msg) reportJsonTR :: HTFJsonObj a => a -> TR () reportJsonTR x = reportLazyBytesTR Info (decodeObj x) data ReportLevel = Debug | Info deriving (Eq,Ord) reportGen :: TestConfig -> ReportLevel -> (Handle -> IO ()) -> TR () reportGen tc level fun = unless (tc_quiet tc && level < Info) $ case tc_output tc of TestOutputHandle h _ -> liftIO (fun h) TestOutputSplitted fp -> do -- split mode: one file for each result to avoid locking on windows ix <- gets ts_index let realFp = fp ++ (show ix) -- just append the index at the end of the file given as output parameter modify (\x -> x { ts_index = ts_index x + 1 }) liftIO $ withFile realFp WriteMode fun
skogsbaer/HTF
Test/Framework/TestReporter.hs
lgpl-2.1
12,583
0
31
3,265
2,865
1,470
1,395
247
6
module ViperVM.Library.FloatMatrixSub ( builtin, function, metaKernel, kernels ) where import Control.Applicative ( (<$>) ) import ViperVM.VirtualPlatform.FunctionalKernel hiding (metaKernel,proto) import ViperVM.VirtualPlatform.MetaObject import ViperVM.VirtualPlatform.Descriptor import ViperVM.VirtualPlatform.MetaKernel hiding (proto,name,kernels) import ViperVM.VirtualPlatform.Objects.Matrix import ViperVM.VirtualPlatform.Object import ViperVM.Platform.KernelParameter import ViperVM.Platform.Kernel import ViperVM.Platform.Peer.KernelPeer import qualified ViperVM.Library.OpenCL.FloatMatrixSub as CL ---------------------------------------- -- Builtin & Function --------------------------------------- builtin :: MakeBuiltin builtin = makeBuiltinIO function function :: IO FunctionalKernel function = FunctionalKernel proto makeParams makeResult <$> metaKernel where proto = Prototype { inputs = [MatrixType,MatrixType], output = MatrixType } makeParams args = do let [a,b] = args c <- allocate (descriptor a) return [a,b,c] makeResult args = last args ---------------------------------------- -- MetaKernel --------------------------------------- metaKernel :: IO MetaKernel metaKernel = MetaKernel name proto conf <$> kernels where name = "FloatMatrixSub" proto = [ Arg ReadOnly "a", Arg ReadOnly "b", Arg WriteOnly "c" ] conf :: [ObjectPeer] -> [KernelParameter] conf objs = params where [MatrixObject ma, MatrixObject mb, MatrixObject mc] = objs params = [WordParam (fromIntegral width), WordParam (fromIntegral height), BufferParam (matrixBuffer ma), WordParam (fromIntegral lda), WordParam (fromIntegral (matrixOffset ma `div` 4)), BufferParam (matrixBuffer mb), WordParam (fromIntegral ldb), WordParam (fromIntegral (matrixOffset mb `div` 4)), BufferParam (matrixBuffer mc), WordParam (fromIntegral ldc), WordParam (fromIntegral (matrixOffset mc `div` 4))] (width, height) = matrixDimensions ma lda = ((matrixWidth ma) * 4 + (matrixPadding ma)) `div` 4 ldb = ((matrixWidth mb) * 4 + (matrixPadding mb)) `div` 4 ldc = ((matrixWidth mc) * 4 + (matrixPadding mc)) `div` 4 ---------------------------------------- -- Kernels --------------------------------------- kernels :: IO [Kernel] kernels = initKernelsIO [ CLKernel <$> CL.kernel ]
hsyl20/HViperVM
lib/ViperVM/Library/FloatMatrixSub.hs
lgpl-3.0
2,695
0
15
691
686
386
300
54
1
module ProjectEuler.A268398Spec (main, spec) where import Test.Hspec import ProjectEuler.A268398 (a268398) main :: IO () main = hspec spec spec :: Spec spec = describe "A268398" $ it "correctly computes the first 20 elements" $ take 20 (map a268398 [1..]) `shouldBe` expectedValue where expectedValue = [1,2,3,7,8,9,10,14,17,18,19,23,24,25,26,42,43,46,47,51]
peterokagey/haskellOEIS
test/ProjectEuler/A268398Spec.hs
apache-2.0
373
0
10
59
160
95
65
10
1
import System.Environment (getArgs) import Control.Arrow (first) import Graphics.Rendering.Chart.Easy hiding (argument) import Graphics.Rendering.Chart.Backend.Diagrams import Data.Time import System.Process import Options.Applicative import Data.Semigroup ((<>)) import Data.Text.Time (parseISODateTime) import qualified Data.TimeSeries as TS -- Command line params data Params = Params { resolution :: String , fileName :: String } deriving (Show) -- Command line parser params :: Parser Params params = Params <$> strOption ( long "time" <> metavar "<Y | M | D | H>" <> help "Resampling resolution." ) <*> argument str (metavar "FILE") -- Convert string representation of time into TimeResolution convertTime :: String -> TS.TimeResolution convertTime "Y" = TS.years 1 convertTime "M" = TS.months 1 convertTime "D" = TS.days 1 convertTime "H" = TS.seconds 3600 -- Main function. Parses command line arguments and calls drawResampled for Time Series processing main :: IO () main = do ps <- execParser opts ts <- TS.loadCSV TS.HasHeader parseISODateTime (fileName ps) drawResampled ts (convertTime (resolution ps)) where opts = info (helper <*> params) ( fullDesc <> progDesc "Resampling time series" ) -- Convert Time series to format expected by Chart library signal :: Num a => TS.Series a -> [(LocalTime, a)] signal ts = map (first (utcToLocalTime utc)) (TS.toList ts) -- Take time series with time resolutions and creates chart with original and modified series drawResampled :: TS.Series Double -> TS.TimeResolution -> IO () drawResampled ts dt = do args <- getArgs let fn = head args let startTime = UTCTime (fromGregorian 1965 1 1) 0 let xs = TS.resample startTime dt ts toFile def "dist/resample.svg" $ do layout_title .= fn setColors [opaque blue, opaque red] plot (line "Original series" [signal ts]) plot (line "Resampled series" [signal xs]) putStrLn "Plot saved to: dist/resample.svg" _ <- createProcess (shell "firefox dist/resample.svg") return ()
klangner/timeseries
examples/Resample.hs
bsd-2-clause
2,108
17
14
430
661
312
349
48
1
{-# LANGUAGE GeneralizedNewtypeDeriving #-} ----------------------------------------------------------------------------- -- | -- Module : Graphics.Hoodle.Render.Util.HitTest -- Copyright : (c) 2011-2013 Ian-Woo Kim -- -- License : BSD3 -- Maintainer : Ian-Woo Kim <[email protected]> -- Stability : experimental -- Portability : GHC -- -- Hit-testing routines -- ----------------------------------------------------------------------------- module Graphics.Hoodle.Render.Util.HitTest where import Control.Applicative import Control.Monad.State import Data.Strict.Tuple -- from hoodle-platform import Data.Hoodle.Simple import Data.Hoodle.BBox -- from this package import Graphics.Hoodle.Render.Type.HitTest import Graphics.Hoodle.Render.Type.Item -------------------------- -- hit test collections -- -------------------------- -- | hit test of whether a point in a bbox -- previously, hitTestBBoxPoint isPointInBBox :: BBox -> (Double,Double) -- ^ point -> Bool isPointInBBox (BBox (ulx,uly) (lrx,lry)) (x,y) = ulx <= x && x <= lrx && uly <= y && y <= lry -- | hit test of whether two lines intersect do2LinesIntersect :: ((Double,Double),(Double,Double)) -- ^ line1 -> ((Double,Double),(Double,Double)) -- ^ line2 -> Bool do2LinesIntersect ((x1,y1),(x2,y2)) ((x3,y3),(x4,y4)) = (x2-xc)*(xc-x1)>0 && (x3-xc)*(xc-x4) >0 where x21 = x2-x1 x43 = x4-x3 y21 = y2-y1 y43 = y4-y3 xc = (x21*x43*(y3-y1)+y21*x43*x1-y43*x21*x3)/(y21*x43-y43*x21) -- | previously, hitTestLineStroke doesLineHitStrk :: ((Double,Double),(Double,Double)) -> Stroke -> Bool doesLineHitStrk line1 str@(Stroke _t _c _w _d) = test (stroke_data str) where test [] = False test ((_:!:_):[]) = False test ((x0:!:y0):(x:!:y):rest) = do2LinesIntersect line1 ((x0,y0),(x,y)) || test ((x:!:y) : rest) doesLineHitStrk line1 (VWStroke _t _c d) = test d where test [] = False test ((_,_,_):[]) = False test ((x0,y0,_):(x,y,z):rest) = do2LinesIntersect line1 ((x0,y0),(x,y)) || test ((x,y,z) : rest) -- | Do two bounding boxes intersect with each other? -- previously, hitTestBBoxBBox do2BBoxIntersect :: BBox -> BBox -> Bool do2BBoxIntersect (BBox (ulx1,uly1) (lrx1,lry1)) (BBox (ulx2,uly2) (lrx2,lry2)) = p1 && p2 where p1 = ulx1 <= ulx2 && ulx2 <= lrx1 || ulx1 <= lrx2 && lrx2 <= lrx1 || ulx2 <= ulx1 && ulx1 <= lrx2 || ulx2 <= lrx1 && lrx1 <= lrx2 p2 = uly1 <= uly2 && uly2 <= lry1 || uly1 <= lry2 && lry2 <= lry1 || uly2 <= uly1 && uly1 <= lry2 || uly2 <= lry1 && lry1 <= lry2 -- | is the second bbox inside the first bbox? -- previously, hitTestInsideBBox isBBox2InBBox1 :: BBox -- ^ 1st bbox -> BBox -- ^ 2nd bbox -> Bool isBBox2InBBox1 b1 (BBox (ulx2,uly2) (lrx2,lry2)) = isPointInBBox b1 (ulx2,uly2) && isPointInBBox b1 (lrx2,lry2) -------------------------------------------------------- -- item filtering functions that results in AlterList -- -------------------------------------------------------- -- | hltFilteredBy_StateT :: (a -> Bool) -- ^ hit test condition -> [a] -- ^ strokes to test -> State Bool (AlterList (NotHitted a) (Hitted a)) hltFilteredBy_StateT test itms = do let (nhit,rest) = break test itms (hit,rest') = break (not.test) rest modify (|| (not.null) hit) if null rest' then return (NotHitted nhit :- Hitted hit :- NotHitted [] :- Empty) else return (NotHitted nhit :- Hitted hit :- hltFilteredBy test rest') -- | highlight strokes filtered by a condition. -- previously mkHitTestAL hltFilteredBy :: (a -> Bool) -- ^ hit test condition -> [a] -- ^ strokes to test -> AlterList (NotHitted a) (Hitted a) hltFilteredBy test is = evalState (hltFilteredBy_StateT test is) False -- | hltHittedByBBox :: (GetBBoxable a) => BBox -- ^ test bounding box -> [a] -- ^ items to test -> AlterList (NotHitted a) (Hitted a) hltHittedByBBox b = hltFilteredBy (do2BBoxIntersect b . getBBox) -- | hltEmbeddedByBBox :: (GetBBoxable a) => BBox -> [a] -> AlterList (NotHitted a) (Hitted a) hltEmbeddedByBBox b = hltFilteredBy (isBBox2InBBox1 b . getBBox) -- | only check if a line and bbox of item overlapped hltHittedByLineRough :: (GetBBoxable a) => ((Double,Double),(Double,Double)) -- ^ line -> [a] -- ^ items to test -> AlterList (NotHitted a) (Hitted a) hltHittedByLineRough (p1,p2) = hltFilteredBy boxhittest where boxhittest s = isPointInBBox (getBBox s) p1 || isPointInBBox (getBBox s) p2 -- | hltItmsHittedByLine_StateT :: ((Double,Double),(Double,Double)) -> [RItem] -> State Bool RItemHitted -- (AlterList (NotHitted RItem) (Hitted RItem)) hltItmsHittedByLine_StateT line = hltFilteredBy_StateT test where test (RItemStroke strk) = (doesLineHitStrk line . bbxed_content) strk test _ = False -- | hltItmsHittedByLineFrmSelected_StateT :: ((Double,Double),(Double,Double)) -> RItemHitted -- AlterList (NotHitted RItem) (Hitted RItem) -> State Bool (AlterList (NotHitted RItem) RItemHitted) hltItmsHittedByLineFrmSelected_StateT _ Empty = error "something is wrong, invariant broken" hltItmsHittedByLineFrmSelected_StateT _ (n:-Empty) = return (n:-Empty) hltItmsHittedByLineFrmSelected_StateT line (n:-h:-rest) = do h' <- hltItmsHittedByLine_StateT line (unHitted h) (n:-) . (h':-) <$> hltItmsHittedByLineFrmSelected_StateT line rest -- | elimHitted :: (GetBBoxable a) => AlterList (NotHitted a) (Hitted a) -> State (Maybe BBox) [a] elimHitted Empty = error "something wrong in elimHitted" elimHitted (n:-Empty) = return (unNotHitted n) elimHitted (n:-h:-rest) = do bbox <- get let bbox2 = getTotalBBox (unHitted h) put (merge bbox bbox2) return . (unNotHitted n ++) =<< elimHitted rest -- | merge :: Maybe BBox -> Maybe BBox -> Maybe BBox merge Nothing Nothing = Nothing merge Nothing (Just b) = Just b merge (Just b) Nothing = Just b merge (Just (BBox (x1,y1) (x2,y2))) (Just (BBox (x3,y3) (x4,y4))) = Just (BBox (min x1 x3, min y1 y3) (max x2 x4,max y2 y4)) -- | getTotalBBox :: (GetBBoxable a) => [a] -> Maybe BBox getTotalBBox = foldl f Nothing where f acc = merge acc . Just . getBBox
wavewave/hoodle-render
src/Graphics/Hoodle/Render/Util/HitTest.hs
bsd-2-clause
6,782
0
21
1,751
2,152
1,174
978
118
5
{-# LANGUAGE DeriveDataTypeable #-} module Gwist.JSON ( Result (..), decodeJSON, readJSON ) where import Data.Typeable import qualified Data.Attoparsec as AP import qualified Data.Aeson as AE import qualified Data.ByteString as BS import qualified Data.ByteString.Lazy as LBS data Result a = Success a | Error String decodeJSON :: AE.FromJSON a => LBS.ByteString -> Result a decodeJSON b = case AP.parse (fmap AE.fromJSON AE.json) (lbsToBs b) of AP.Done _ (AE.Success g) -> Success g AP.Done _ (AE.Error msg) -> Error msg AP.Fail _ _ msg -> Error msg AP.Partial _ -> Error "JSON is too short" where lbsToBs = BS.concat . LBS.toChunks readJSON :: AE.FromJSON a => LBS.ByteString -> IO a readJSON b = case decodeJSON b of Success g -> return g Error msg -> error msg
shugo/gwist
Gwist/JSON.hs
bsd-2-clause
816
0
11
176
292
153
139
24
4
module Propellor.Property.Dns ( module Propellor.Types.Dns, primary, signedPrimary, secondary, secondaryFor, mkSOA, writeZoneFile, nextSerialNumber, adjustSerialNumber, serialNumberOffset, WarningMessage, genZone, ) where import Propellor.Base import Propellor.Types.Dns import Propellor.Types.Info import Propellor.Property.File import qualified Propellor.Property.Apt as Apt import qualified Propellor.Property.Ssh as Ssh import qualified Propellor.Property.Service as Service import Propellor.Property.Scheduled import Propellor.Property.DnsSec import Utility.Applicative import qualified Data.Map as M import qualified Data.Set as S import Data.List -- | Primary dns server for a domain, using bind. -- -- Currently, this only configures bind to serve forward DNS, not reverse DNS. -- -- Most of the content of the zone file is configured by setting properties -- of hosts. For example, -- -- > host "foo.example.com" -- > & ipv4 "192.168.1.1" -- > & alias "mail.exmaple.com" -- -- Will cause that hostmame and its alias to appear in the zone file, -- with the configured IP address. -- -- Also, if a host has a ssh public key configured, a SSHFP record will -- be automatically generated for it. -- -- The [(BindDomain, Record)] list can be used for additional records -- that cannot be configured elsewhere. This often includes NS records, -- TXT records and perhaps CNAMEs pointing at hosts that propellor does -- not control. -- -- The primary server is configured to only allow zone transfers to -- secondary dns servers. These are determined in two ways: -- -- 1. By looking at the properties of other hosts, to find hosts that -- are configured as the secondary dns server. -- -- 2. By looking for NS Records in the passed list of records. -- -- In either case, the secondary dns server Host should have an ipv4 and/or -- ipv6 property defined. primary :: [Host] -> Domain -> SOA -> [(BindDomain, Record)] -> RevertableProperty (HasInfo + DebianLike) DebianLike primary hosts domain soa rs = setup <!> cleanup where setup = setupPrimary zonefile id hosts domain soa rs `onChange` Service.reloaded "bind9" cleanup = cleanupPrimary zonefile domain `onChange` Service.reloaded "bind9" zonefile = "/etc/bind/propellor/db." ++ domain setupPrimary :: FilePath -> (FilePath -> FilePath) -> [Host] -> Domain -> SOA -> [(BindDomain, Record)] -> Property (HasInfo + DebianLike) setupPrimary zonefile mknamedconffile hosts domain soa rs = withwarnings baseprop `requires` servingZones where hostmap = hostMap hosts -- Known hosts with hostname located in the domain. indomain = M.elems $ M.filterWithKey (\hn _ -> inDomain domain $ AbsDomain $ hn) hostmap (partialzone, zonewarnings) = genZone indomain hostmap domain soa baseprop = primaryprop `setInfoProperty` (toInfo (addNamedConf conf)) primaryprop :: Property DebianLike primaryprop = property ("dns primary for " ++ domain) $ do sshfps <- concat <$> mapM (genSSHFP domain) (M.elems hostmap) let zone = partialzone { zHosts = zHosts partialzone ++ rs ++ sshfps } ifM (liftIO $ needupdate zone) ( makeChange $ writeZoneFile zone zonefile , noChange ) withwarnings p = adjustPropertySatisfy p $ \a -> do mapM_ warningMessage $ zonewarnings ++ secondarywarnings a conf = NamedConf { confDomain = domain , confDnsServerType = Master , confFile = mknamedconffile zonefile , confMasters = [] , confAllowTransfer = nub $ concatMap (`hostAddresses` hosts) $ secondaries ++ nssecondaries , confLines = [] } secondaries = otherServers Secondary hosts domain secondarywarnings = map (\h -> "No IP address defined for DNS seconary " ++ h) $ filter (\h -> null (hostAddresses h hosts)) secondaries nssecondaries = mapMaybe (domainHostName <=< getNS) rootRecords rootRecords = map snd $ filter (\(d, _r) -> d == RootDomain || d == AbsDomain domain) rs needupdate zone = do v <- readZonePropellorFile zonefile return $ case v of Nothing -> True Just oldzone -> -- compare everything except serial let oldserial = sSerial (zSOA oldzone) z = zone { zSOA = (zSOA zone) { sSerial = oldserial } } in z /= oldzone || oldserial < sSerial (zSOA zone) cleanupPrimary :: FilePath -> Domain -> Property DebianLike cleanupPrimary zonefile domain = check (doesFileExist zonefile) $ go `requires` namedConfWritten where desc = "removed dns primary for " ++ domain go :: Property DebianLike go = property desc (makeChange $ removeZoneFile zonefile) -- | Primary dns server for a domain, secured with DNSSEC. -- -- This is like `primary`, except the resulting zone -- file is signed. -- The Zone Signing Key (ZSK) and Key Signing Key (KSK) -- used in signing it are taken from the PrivData. -- -- As a side effect of signing the zone, a -- </var/cache/bind/dsset-domain.> -- file will be created. This file contains the DS records -- which need to be communicated to your domain registrar -- to make DNSSEC be used for your domain. Doing so is outside -- the scope of propellor (currently). See for example the tutorial -- <https://www.digitalocean.com/community/tutorials/how-to-setup-dnssec-on-an-authoritative-bind-dns-server--2> -- -- The 'Recurrance' controls how frequently the signature -- should be regenerated, using a new random salt, to prevent -- zone walking attacks. `Weekly Nothing` is a reasonable choice. -- -- To transition from 'primary' to 'signedPrimary', you can revert -- the 'primary' property, and add this property. -- -- Note that DNSSEC zone files use a serial number based on the unix epoch. -- This is different from the serial number used by 'primary', so if you -- want to later disable DNSSEC you will need to adjust the serial number -- passed to mkSOA to ensure it is larger. signedPrimary :: Recurrance -> [Host] -> Domain -> SOA -> [(BindDomain, Record)] -> RevertableProperty (HasInfo + DebianLike) DebianLike signedPrimary recurrance hosts domain soa rs = setup <!> cleanup where setup = combineProperties ("dns primary for " ++ domain ++ " (signed)") (props & setupPrimary zonefile signedZoneFile hosts domain soa rs' & zoneSigned domain zonefile & forceZoneSigned domain zonefile `period` recurrance ) `onChange` Service.reloaded "bind9" cleanup = cleanupPrimary zonefile domain `onChange` revert (zoneSigned domain zonefile) `onChange` Service.reloaded "bind9" -- Include the public keys into the zone file. rs' = include PubKSK : include PubZSK : rs include k = (RootDomain, INCLUDE (keyFn domain k)) -- Put DNSSEC zone files in a different directory than is used for -- the regular ones. This allows 'primary' to be reverted and -- 'signedPrimary' enabled, without the reverted property stomping -- on the new one's settings. zonefile = "/etc/bind/propellor/dnssec/db." ++ domain -- | Secondary dns server for a domain. -- -- The primary server is determined by looking at the properties of other -- hosts to find which one is configured as the primary. -- -- Note that if a host is declared to be a primary and a secondary dns -- server for the same domain, the primary server config always wins. secondary :: [Host] -> Domain -> RevertableProperty (HasInfo + DebianLike) DebianLike secondary hosts domain = secondaryFor (otherServers Master hosts domain) hosts domain -- | This variant is useful if the primary server does not have its DNS -- configured via propellor. secondaryFor :: [HostName] -> [Host] -> Domain -> RevertableProperty (HasInfo + DebianLike) DebianLike secondaryFor masters hosts domain = setup <!> cleanup where setup = pureInfoProperty desc (addNamedConf conf) `requires` servingZones cleanup = namedConfWritten desc = "dns secondary for " ++ domain conf = NamedConf { confDomain = domain , confDnsServerType = Secondary , confFile = "db." ++ domain , confMasters = concatMap (`hostAddresses` hosts) masters , confAllowTransfer = [] , confLines = [] } otherServers :: DnsServerType -> [Host] -> Domain -> [HostName] otherServers wantedtype hosts domain = M.keys $ M.filter wanted $ hostMap hosts where wanted h = case M.lookup domain (fromNamedConfMap $ fromInfo $ hostInfo h) of Nothing -> False Just conf -> confDnsServerType conf == wantedtype && confDomain conf == domain -- | Rewrites the whole named.conf.local file to serve the zones -- configured by `primary` and `secondary`, and ensures that bind9 is -- running. servingZones :: Property DebianLike servingZones = namedConfWritten `onChange` Service.reloaded "bind9" `requires` Apt.serviceInstalledRunning "bind9" namedConfWritten :: Property DebianLike namedConfWritten = property' "named.conf configured" $ \w -> do zs <- getNamedConf ensureProperty w $ hasContent namedConfFile $ concatMap confStanza $ M.elems zs confStanza :: NamedConf -> [Line] confStanza c = [ "// automatically generated by propellor" , "zone \"" ++ confDomain c ++ "\" {" , cfgline "type" (if confDnsServerType c == Master then "master" else "slave") , cfgline "file" ("\"" ++ confFile c ++ "\"") ] ++ mastersblock ++ allowtransferblock ++ (map (\l -> "\t" ++ l ++ ";") (confLines c)) ++ [ "};" , "" ] where cfgline f v = "\t" ++ f ++ " " ++ v ++ ";" ipblock name l = [ "\t" ++ name ++ " {" ] ++ (map (\ip -> "\t\t" ++ val ip ++ ";") l) ++ [ "\t};" ] mastersblock | null (confMasters c) = [] | otherwise = ipblock "masters" (confMasters c) -- an empty block prohibits any transfers allowtransferblock = ipblock "allow-transfer" (confAllowTransfer c) namedConfFile :: FilePath namedConfFile = "/etc/bind/named.conf.local" -- | Generates a SOA with some fairly sane numbers in it. -- -- The Domain is the domain to use in the SOA record. Typically -- something like ns1.example.com. So, not the domain that this is the SOA -- record for. -- -- The SerialNumber can be whatever serial number was used by the domain -- before propellor started managing it. Or 0 if the domain has only ever -- been managed by propellor. -- -- You do not need to increment the SerialNumber when making changes! -- Propellor will automatically add the number of commits in the git -- repository to the SerialNumber. mkSOA :: Domain -> SerialNumber -> SOA mkSOA d sn = SOA { sDomain = AbsDomain d , sSerial = sn , sRefresh = hours 4 , sRetry = hours 1 , sExpire = 2419200 -- 4 weeks , sNegativeCacheTTL = hours 8 } where hours n = n * 60 * 60 dValue :: BindDomain -> String dValue (RelDomain d) = d dValue (AbsDomain d) = d ++ "." dValue (RootDomain) = "@" rField :: Record -> Maybe String rField (Address (IPv4 _)) = Just "A" rField (Address (IPv6 _)) = Just "AAAA" rField (CNAME _) = Just "CNAME" rField (MX _ _) = Just "MX" rField (NS _) = Just "NS" rField (TXT _) = Just "TXT" rField (SRV _ _ _ _) = Just "SRV" rField (SSHFP _ _ _) = Just "SSHFP" rField (INCLUDE _) = Just "$INCLUDE" rField (PTR _) = Nothing rValue :: Record -> Maybe String rValue (Address (IPv4 addr)) = Just addr rValue (Address (IPv6 addr)) = Just addr rValue (CNAME d) = Just $ dValue d rValue (MX pri d) = Just $ val pri ++ " " ++ dValue d rValue (NS d) = Just $ dValue d rValue (SRV priority weight port target) = Just $ unwords [ val priority , val weight , val port , dValue target ] rValue (SSHFP x y s) = Just $ unwords [ val x , val y , s ] rValue (INCLUDE f) = Just f rValue (TXT s) = Just $ [q] ++ filter (/= q) s ++ [q] where q = '"' rValue (PTR _) = Nothing -- | Adjusts the serial number of the zone to always be larger -- than the serial number in the Zone record, -- and always be larger than the passed SerialNumber. nextSerialNumber :: Zone -> SerialNumber -> Zone nextSerialNumber z serial = adjustSerialNumber z $ \sn -> succ $ max sn serial adjustSerialNumber :: Zone -> (SerialNumber -> SerialNumber) -> Zone adjustSerialNumber (Zone d soa l) f = Zone d soa' l where soa' = soa { sSerial = f (sSerial soa) } -- | Count the number of git commits made to the current branch. serialNumberOffset :: IO SerialNumber serialNumberOffset = fromIntegral . length . lines <$> readProcess "git" ["log", "--pretty=%H"] -- | Write a Zone out to a to a file. -- -- The serial number in the Zone automatically has the serialNumberOffset -- added to it. Also, just in case, the old serial number used in the zone -- file is checked, and if it is somehow larger, its succ is used. writeZoneFile :: Zone -> FilePath -> IO () writeZoneFile z f = do oldserial <- oldZoneFileSerialNumber f offset <- serialNumberOffset let z' = nextSerialNumber (adjustSerialNumber z (+ offset)) oldserial createDirectoryIfMissing True (takeDirectory f) writeFile f (genZoneFile z') writeZonePropellorFile f z' removeZoneFile :: FilePath -> IO () removeZoneFile f = do nukeFile f nukeFile (zonePropellorFile f) -- | Next to the zone file, is a ".propellor" file, which contains -- the serialized Zone. This saves the bother of parsing -- the horrible bind zone file format. zonePropellorFile :: FilePath -> FilePath zonePropellorFile f = f ++ ".propellor" oldZoneFileSerialNumber :: FilePath -> IO SerialNumber oldZoneFileSerialNumber = maybe 0 (sSerial . zSOA) <$$> readZonePropellorFile writeZonePropellorFile :: FilePath -> Zone -> IO () writeZonePropellorFile f z = writeFile (zonePropellorFile f) (show z) readZonePropellorFile :: FilePath -> IO (Maybe Zone) readZonePropellorFile f = catchDefaultIO Nothing $ readish <$> readFileStrict (zonePropellorFile f) -- | Generating a zone file. genZoneFile :: Zone -> String genZoneFile (Zone zdomain soa rs) = unlines $ header : genSOA soa ++ mapMaybe (genRecord zdomain) rs where header = com $ "BIND zone file for " ++ zdomain ++ ". Generated by propellor, do not edit." genRecord :: Domain -> (BindDomain, Record) -> Maybe String genRecord zdomain (domain, record) = case (rField record, rValue record) of (Nothing, _) -> Nothing (_, Nothing) -> Nothing (Just rfield, Just rvalue) -> Just $ intercalate "\t" $ case record of INCLUDE _ -> [ rfield, rvalue ] _ -> [ domainHost zdomain domain , "IN" , rfield , rvalue ] genSOA :: SOA -> [String] genSOA soa = -- "@ IN SOA ns1.example.com. root (" [ intercalate "\t" [ dValue RootDomain , "IN" , "SOA" , dValue (sDomain soa) , "root" , "(" ] , headerline sSerial "Serial" , headerline sRefresh "Refresh" , headerline sRetry "Retry" , headerline sExpire "Expire" , headerline sNegativeCacheTTL "Negative Cache TTL" , inheader ")" ] where headerline r comment = inheader $ show (r soa) ++ "\t\t" ++ com comment inheader l = "\t\t\t" ++ l -- | Comment line in a zone file. com :: String -> String com s = "; " ++ s type WarningMessage = String -- | Generates a Zone for a particular Domain from the DNS properies of all -- hosts that propellor knows about that are in that Domain. -- -- Does not include SSHFP records. genZone :: [Host] -> M.Map HostName Host -> Domain -> SOA -> (Zone, [WarningMessage]) genZone inzdomain hostmap zdomain soa = let (warnings, zhosts) = partitionEithers $ concatMap concat [ map hostips inzdomain , map hostrecords inzdomain , map addcnames (M.elems hostmap) ] in (Zone zdomain soa (simplify zhosts), warnings) where -- Each host with a hostname located in the zdomain -- should have 1 or more IPAddrs in its Info. -- -- If a host lacks any IPAddr, it's probably a misconfiguration, -- so warn. hostips :: Host -> [Either WarningMessage (BindDomain, Record)] hostips h | null l = [Left $ "no IP address defined for host " ++ hostName h] | otherwise = map Right l where info = hostInfo h l = zip (repeat $ AbsDomain $ hostName h) (map Address $ getAddresses info) -- Any host, whether its hostname is in the zdomain or not, -- may have cnames which are in the zdomain. The cname may even be -- the same as the root of the zdomain, which is a nice way to -- specify IP addresses for a SOA record. -- -- Add Records for those.. But not actually, usually, cnames! -- Why not? Well, using cnames doesn't allow doing some things, -- including MX and round robin DNS, and certianly CNAMES -- shouldn't be used in SOA records. -- -- We typically know the host's IPAddrs anyway. -- So we can just use the IPAddrs. addcnames :: Host -> [Either WarningMessage (BindDomain, Record)] addcnames h = concatMap gen $ filter (inDomain zdomain) $ mapMaybe getCNAME $ S.toList $ fromDnsInfo $ fromInfo info where info = hostInfo h gen c = case getAddresses info of [] -> [ret (CNAME c)] l -> map (ret . Address) l where ret record = Right (c, record) -- Adds any other DNS records for a host located in the zdomain. hostrecords :: Host -> [Either WarningMessage (BindDomain, Record)] hostrecords h = map Right l where info = hostInfo h l = zip (repeat $ AbsDomain $ hostName h) (S.toList $ S.filter (\r -> isNothing (getIPAddr r) && isNothing (getCNAME r)) (fromDnsInfo $ fromInfo info)) -- Simplifies the list of hosts. Remove duplicate entries. -- Also, filter out any CHAMES where the same domain has an -- IP address, since that's not legal. simplify :: [(BindDomain, Record)] -> [(BindDomain, Record)] simplify l = nub $ filter (not . dupcname ) l where dupcname (d, CNAME _) | any (matchingaddr d) l = True dupcname _ = False matchingaddr d (d', (Address _)) | d == d' = True matchingaddr _ _ = False inDomain :: Domain -> BindDomain -> Bool inDomain domain (AbsDomain d) = domain == d || ('.':domain) `isSuffixOf` d inDomain _ _ = False -- can't tell, so assume not -- | Gets the hostname of the second domain, relative to the first domain, -- suitable for using in a zone file. domainHost :: Domain -> BindDomain -> String domainHost _ (RelDomain d) = d domainHost _ RootDomain = "@" domainHost base (AbsDomain d) | dotbase `isSuffixOf` d = take (length d - length dotbase) d | base == d = "@" | otherwise = d where dotbase = '.':base addNamedConf :: NamedConf -> NamedConfMap addNamedConf conf = NamedConfMap (M.singleton domain conf) where domain = confDomain conf getNamedConf :: Propellor (M.Map Domain NamedConf) getNamedConf = asks $ fromNamedConfMap . fromInfo . hostInfo -- | Generates SSHFP records for hosts in the domain (or with CNAMES -- in the domain) that have configured ssh public keys. -- -- This is done using ssh-keygen, so sadly needs IO. genSSHFP :: Domain -> Host -> Propellor [(BindDomain, Record)] genSSHFP domain h = concatMap mk . concat <$> (gen =<< get) where get = fromHost [h] hostname Ssh.getHostPubKey gen = liftIO . mapM genSSHFP' . M.elems . fromMaybe M.empty mk r = mapMaybe (\d -> if inDomain domain d then Just (d, r) else Nothing) (AbsDomain hostname : cnames) cnames = mapMaybe getCNAME $ S.toList $ fromDnsInfo $ fromInfo info hostname = hostName h info = hostInfo h genSSHFP' :: String -> IO [Record] genSSHFP' pubkey = withTmpFile "sshfp" $ \tmp tmph -> do hPutStrLn tmph pubkey hClose tmph s <- catchDefaultIO "" $ readProcess "ssh-keygen" ["-r", "dummy", "-f", tmp] return $ mapMaybe (parse . words) $ lines s where parse ("dummy":"IN":"SSHFP":x:y:s:[]) = do x' <- readish x y' <- readish y return $ SSHFP x' y' s parse _ = Nothing
ArchiveTeam/glowing-computing-machine
src/Propellor/Property/Dns.hs
bsd-2-clause
19,165
748
20
3,628
4,997
2,719
2,278
-1
-1
{-# LANGUAGE MultiParamTypeClasses, FlexibleContexts, FlexibleInstances, PatternGuards #-} ----------------------------------------------------------------------------- -- | -- Module : XMonad.Layout.ComboP -- Description : Combine multiple layouts and specify where to put new windows. -- Copyright : (c) Konstantin Sobolev <[email protected]> -- License : BSD-style (see LICENSE) -- -- Maintainer : Konstantin Sobolev <[email protected]> -- Stability : unstable -- Portability : unportable -- -- A layout that combines multiple layouts and allows to specify where to put -- new windows. -- ----------------------------------------------------------------------------- module XMonad.Layout.ComboP ( -- * Usage -- $usage combineTwoP, CombineTwoP, SwapWindow(..), PartitionWins(..), Property(..) ) where import XMonad.Prelude import XMonad hiding (focus) import XMonad.StackSet ( Workspace (..), Stack(..) ) import XMonad.Layout.WindowNavigation import XMonad.Util.WindowProperties import qualified XMonad.StackSet as W -- $usage -- You can use this module with the following in your @~\/.xmonad\/xmonad.hs@: -- -- > import XMonad.Layout.ComboP -- -- and add something like -- -- > combineTwoP (TwoPane 0.03 0.5) (tabbed shrinkText defaultTConf) (tabbed shrinkText defaultTConf) (ClassName "Firefox") -- -- to your layouts. This way all windows with class = \"Firefox\" will always go -- to the left pane, all others - to the right. -- -- For more detailed instructions on editing the layoutHook see: -- -- "XMonad.Doc.Extending#Editing_the_layout_hook" -- -- 'combineTwoP' is a simple layout combinator based on 'combineTwo' from Combo, with -- addition of a 'Property' which tells where to put new windows. Windows mathing -- the property will go into the first part, all others will go into the second -- part. It supports @Move@ messages as 'combineTwo' does, but it also introduces -- 'SwapWindow' message which sends focused window to the other part. It is -- required because @Move@ commands don't work when one of the parts is empty. -- To use it, import \"XMonad.Layout.WindowNavigation\", and add the following key -- bindings (or something similar): -- -- > , ((modm .|. controlMask .|. shiftMask, xK_Right), sendMessage $ Move R) -- > , ((modm .|. controlMask .|. shiftMask, xK_Left ), sendMessage $ Move L) -- > , ((modm .|. controlMask .|. shiftMask, xK_Up ), sendMessage $ Move U) -- > , ((modm .|. controlMask .|. shiftMask, xK_Down ), sendMessage $ Move D) -- > , ((modm .|. controlMask .|. shiftMask, xK_s ), sendMessage $ SwapWindow) -- -- For detailed instruction on editing the key binding see -- "XMonad.Doc.Extending#Editing_key_bindings". data SwapWindow = SwapWindow -- ^ Swap window between panes | SwapWindowN Int -- ^ Swap window between panes in the N-th nested ComboP. @SwapWindowN 0@ equals to SwapWindow deriving (Read, Show) instance Message SwapWindow data PartitionWins = PartitionWins -- ^ Reset the layout and -- partition all windows into the -- correct sub-layout. Useful for -- when window properties have -- changed and you want ComboP to -- update which layout a window -- belongs to. deriving (Read, Show) instance Message PartitionWins data CombineTwoP l l1 l2 a = C2P [a] [a] [a] l (l1 a) (l2 a) Property deriving (Read, Show) combineTwoP :: (LayoutClass super(), LayoutClass l1 Window, LayoutClass l2 Window) => super () -> l1 Window -> l2 Window -> Property -> CombineTwoP (super ()) l1 l2 Window combineTwoP = C2P [] [] [] instance (LayoutClass l (), LayoutClass l1 Window, LayoutClass l2 Window) => LayoutClass (CombineTwoP (l ()) l1 l2) Window where doLayout (C2P f w1 w2 super l1 l2 prop) rinput s = let origws = W.integrate s -- passed in windows w1c = origws `intersect` w1 -- current windows in the first pane w2c = origws `intersect` w2 -- current windows in the second pane new = origws \\ (w1c ++ w2c) -- new windows superstack = Just Stack { focus=(), up=[], down=[()] } f' = focus s:delete (focus s) f -- list of focused windows, contains 2 elements at most in do matching <- hasProperty prop `filterM` new -- new windows matching predecate let w1' = w1c ++ matching -- updated first pane windows w2' = w2c ++ (new \\ matching) -- updated second pane windows s1 = differentiate f' w1' -- first pane stack s2 = differentiate f' w2' -- second pane stack ([((),r1),((),r2)], msuper') <- runLayout (Workspace "" super superstack) rinput (wrs1, ml1') <- runLayout (Workspace "" l1 s1) r1 (wrs2, ml2') <- runLayout (Workspace "" l2 s2) r2 return (wrs1++wrs2, Just $ C2P f' w1' w2' (fromMaybe super msuper') (fromMaybe l1 ml1') (fromMaybe l2 ml2') prop) handleMessage us@(C2P f ws1 ws2 super l1 l2 prop) m | Just PartitionWins <- fromMessage m = return . Just $ C2P [] [] [] super l1 l2 prop | Just SwapWindow <- fromMessage m = swap us | Just (SwapWindowN 0) <- fromMessage m = swap us | Just (SwapWindowN n) <- fromMessage m = forwardToFocused us $ SomeMessage $ SwapWindowN $ n-1 | Just (MoveWindowToWindow w1 w2) <- fromMessage m, w1 `elem` ws1, w2 `elem` ws2 = return $ Just $ C2P f (delete w1 ws1) (w1:ws2) super l1 l2 prop | Just (MoveWindowToWindow w1 w2) <- fromMessage m, w1 `elem` ws2, w2 `elem` ws1 = return $ Just $ C2P f (w1:ws1) (delete w1 ws2) super l1 l2 prop | otherwise = do ml1' <- handleMessage l1 m ml2' <- handleMessage l2 m msuper' <- handleMessage super m if isJust msuper' || isJust ml1' || isJust ml2' then return $ Just $ C2P f ws1 ws2 (fromMaybe super msuper') (fromMaybe l1 ml1') (fromMaybe l2 ml2') prop else return Nothing description (C2P _ _ _ super l1 l2 prop) = "combining " ++ description l1 ++ " and "++ description l2 ++ " with " ++ description super ++ " using "++ show prop -- send focused window to the other pane. Does nothing if we don't -- own the focused window swap :: (LayoutClass s a, LayoutClass l1 Window, LayoutClass l2 Window) => CombineTwoP (s a) l1 l2 Window -> X (Maybe (CombineTwoP (s a) l1 l2 Window)) swap (C2P f ws1 ws2 super l1 l2 prop) = do mst <- gets (W.stack . W.workspace . W.current . windowset) let (ws1', ws2') = case mst of Nothing -> (ws1, ws2) Just st -> if foc `elem` ws1 then (foc `delete` ws1, foc:ws2) else if foc `elem` ws2 then (foc:ws1, foc `delete` ws2) else (ws1, ws2) where foc = W.focus st if (ws1,ws2) == (ws1',ws2') then return Nothing else return $ Just $ C2P f ws1' ws2' super l1 l2 prop -- forwards the message to the sublayout which contains the focused window forwardToFocused :: (LayoutClass l1 Window, LayoutClass l2 Window, LayoutClass s a) => CombineTwoP (s a) l1 l2 Window -> SomeMessage -> X (Maybe (CombineTwoP (s a) l1 l2 Window)) forwardToFocused (C2P f ws1 ws2 super l1 l2 prop) m = do ml1 <- forwardIfFocused l1 ws1 m ml2 <- forwardIfFocused l2 ws2 m ms <- if isJust ml1 || isJust ml2 then return Nothing else handleMessage super m if isJust ml1 || isJust ml2 || isJust ms then return $ Just $ C2P f ws1 ws2 (fromMaybe super ms) (fromMaybe l1 ml1) (fromMaybe l2 ml2) prop else return Nothing -- forwards message m to layout l if focused window is among w forwardIfFocused :: (LayoutClass l Window) => l Window -> [Window] -> SomeMessage -> X (Maybe (l Window)) forwardIfFocused l w m = do mst <- gets (W.stack . W.workspace . W.current . windowset) maybe (return Nothing) send mst where send st = if W.focus st `elem` w then handleMessage l m else return Nothing -- code from CombineTwo -- given two sets of zs and xs takes the first z from zs that also belongs to xs -- and turns xs into a stack with z being current element. Acts as -- StackSet.differentiate if zs and xs don't intersect differentiate :: Eq q => [q] -> [q] -> Maybe (Stack q) differentiate (z:zs) xs | z `elem` xs = Just $ Stack { focus=z , up = reverse $ takeWhile (/=z) xs , down = tail $ dropWhile (/=z) xs } | otherwise = differentiate zs xs differentiate [] xs = W.differentiate xs -- vim:ts=4:shiftwidth=4:softtabstop=4:expandtab:foldlevel=20:
xmonad/xmonad-contrib
XMonad/Layout/ComboP.hs
bsd-3-clause
9,693
0
16
3,123
2,184
1,156
1,028
106
5
{-# LANGUAGE Arrows #-} module Arrow where import Control.Arrow import Control.Arrow.Operations integral :: ArrowCircuit a => Int -> a Double Double integral rate = proc x -> do rec let i' = i + x * dt i <- delay 0 -< i' returnA -< i where dt = 1 / fromIntegral rate
svenkeidel/stream-bench
src/Arrow.hs
bsd-3-clause
285
1
14
71
107
53
54
10
1
----------------------------------------------------------------------------- -- | -- Module : Graphics.UI.Toy.Gtk.Prelude -- Copyright : (c) 2012 Michael Sloan -- License : BSD-style (see the LICENSE file) -- -- Maintainer : Michael Sloan <[email protected]> -- Stability : experimental -- Portability : GHC only -- -- This module re-exports the "Diagrams.Prelude", along with all of the -- exported modules in toy-diagrams, toy-gtk, and toy-gtk-diagrams. -- -- While importing this module should be discouraged in \"real\" code, it's -- convenient for \"toy\" code. -- ----------------------------------------------------------------------------- module Graphics.UI.Toy.Gtk.Prelude ( module Diagrams.Prelude , module Graphics.UI.Toy.Gtk , module Graphics.UI.Toy.Gtk.Diagrams , module Graphics.UI.Toy.Gtk.Dict , module Graphics.UI.Toy.Gtk.Text , module Graphics.UI.Toy.Button , module Graphics.UI.Toy.Diagrams , module Graphics.UI.Toy.Dict , module Graphics.UI.Toy.Draggable -- , module Graphics.UI.Toy.Slider , module Graphics.UI.Toy.Transformed , Cairo ) where import Diagrams.Backend.Cairo ( Cairo ) import Diagrams.Prelude import Graphics.UI.Toy.Gtk import Graphics.UI.Toy.Gtk.Diagrams import Graphics.UI.Toy.Gtk.Dict import Graphics.UI.Toy.Gtk.Text import Graphics.UI.Toy.Button import Graphics.UI.Toy.Diagrams import Graphics.UI.Toy.Dict import Graphics.UI.Toy.Draggable -- import Graphics.UI.Toy.Slider import Graphics.UI.Toy.Transformed
mgsloan/toy-gtk-diagrams
src/Graphics/UI/Toy/Gtk/Prelude.hs
bsd-3-clause
1,493
0
5
193
196
148
48
23
0
-- Copyright (c) 2016-present, Facebook, Inc. -- All rights reserved. -- -- This source code is licensed under the BSD-style license found in the -- LICENSE file in the root directory of this source tree. module Duckling.Duration.SV.Tests ( tests ) where import Prelude import Data.String import Test.Tasty import Duckling.Dimensions.Types import Duckling.Duration.SV.Corpus import Duckling.Testing.Asserts tests :: TestTree tests = testGroup "SV Tests" [ makeCorpusTest [Seal Duration] corpus ]
facebookincubator/duckling
tests/Duckling/Duration/SV/Tests.hs
bsd-3-clause
509
0
9
80
79
50
29
11
1
{-# LANGUAGE DeriveFoldable #-} {-# LANGUAGE DeriveFunctor #-} {-# LANGUAGE DeriveGeneric #-} {-# LANGUAGE FlexibleInstances #-} {-# LANGUAGE OverloadedStrings #-} module Language.SimplePi.Types ( Position (..) , dummyPos , Statement (..) , Expr , Ident (..) , Binding (..) , ExprF (..) , getPosition , Identity (..) , Fix (..) ) where import Data.Functor.Foldable (Fix (..)) import Data.Functor.Identity import Data.Text (Text) import qualified Data.Text.Lazy as Lazy import Text.PrettyPrint.Leijen.Text import GHC.Generics -- Statement data Statement = Load FilePath | Parameter Ident Expr {- <var> : <expr> -} | Definition Ident Expr {- <var> = <expr> -} | Check Expr | Eval Expr deriving (Show, Eq, Generic) -- Expression type Expr = Fix ExprF newtype Ident = Ident Text deriving (Show, Eq, Ord, Generic) data Binding f e = Binding Ident (f e) deriving (Show, Eq, Foldable, Functor, Generic) data ExprF e = Lambda Position [Binding Maybe e] e | Pi Position (Binding Identity e) e | Arrow Position e e [e] | App Position e e [e] | Var Position Ident | Universe Position Integer deriving (Show, Eq, Functor, Foldable, Generic) getPosition :: Expr -> Position getPosition (Fix e) = case e of (Lambda pos _ _) -> pos (Pi pos _ _) -> pos (Arrow pos _ _ _) -> pos (App pos _ _ _) -> pos (Var pos _) -> pos (Universe pos _) -> pos -- | File position data Position = Position { posFileName :: !Text , posLine :: {-# UNPACK #-} !Int , posColumn :: {-# UNPACK #-} !Int } deriving (Show, Ord, Eq) dummyPos :: Position dummyPos = Position "<no location information>" 1 0 instance Pretty Position where pretty (Position fn line col) = text (Lazy.fromStrict fn) <> colon <> int line <> colon <> int col
esmolanka/simple-pi
src/Language/SimplePi/Types.hs
bsd-3-clause
1,871
0
13
470
605
345
260
63
6
-- | Transforms MultiCoreStatus' into diagrams module Graphics.Diagrams.Transformations.MultiCoreStatus2Diagram ( transformStatus ) where -- External imports import Data.Maybe import Data.History -- Internal imports import Config.Preferences import Config.Config import Graphics.Diagrams.Simple.Diagram import Graphics.Diagrams.MultiCoreStatus import Graphics.Diagrams.Types (Name) import Model.SystemStatus -- | Transform a multicore status into a diagram transformStatus :: Config -> SystemStatus -> Diagram transformStatus cfg (SystemStatus mhist s) = Diagram ps' ms' where ps' = mapMaybe (transformProcessingUnit cfg s) ps ms' = map transformMessage ms (MultiCoreStatus ps ms) = historyPresent mhist -- | Transforms a processing unit into a group box transformProcessingUnit :: Config -> [Name] -> ProcessingUnit -> Maybe Box transformProcessingUnit _ _ (ProcessingUnit _ _ UnitIgnored) = Nothing transformProcessingUnit cfg sel (ProcessingUnit n cs e) = Just $ GroupBox n cs' col expand where cs' = if expand then map (transformRunningElement cfg sel') cs else [] col = processingUnitColor cfg (sel == [n]) sel' = if not (null sel) && head sel == n then tail sel else [] expand = e == UnitExpanded -- | Transforms a running element (component, application) into a box transformRunningElement :: Config -> [Name] -> RunningElement -> Box transformRunningElement cfg ns (Component n k s _ _) = Box n k (runningElementColor cfg (ns == [n]) s) -- | Transforms a message into an arrow transformMessage :: Message -> Arrow transformMessage m = Arrow (sender m) (receiver m)
ivanperez-keera/SoOSiM-ui
src/Graphics/Diagrams/Transformations/MultiCoreStatus2Diagram.hs
bsd-3-clause
1,641
0
12
292
436
234
202
28
3
module Main where import Data.AhoCorasick main :: IO () main = do let acm = construct ["he", "she", "his", "hers"] print $ run acm "ushers" let acm' = constructWithValues [("he", 1.2), ("she", 3.4), ("his", 5.6), ("hers", 7.8)] print $ run acm' "ushers"
yuttie/ac-machine
examples/Simple.hs
bsd-3-clause
273
0
12
61
120
66
54
8
1
-- {-# LANGUAGE EmptyDataDecls #-} {-# LANGUAGE ExplicitForAll #-} --{-# LANGUAGE FlexibleContexts #-} {-# LANGUAGE FlexibleInstances #-} -- {-# LANGUAGE FunctionalDependencies #-} {-# LANGUAGE MultiParamTypeClasses #-} --{-# LANGUAGE PartialTypeSignatures #-} {-# LANGUAGE RankNTypes #-} --{-# LANGUAGE RebindableSyntax #-} --{-# LANGUAGE ScopedTypeVariables #-} {-# LANGUAGE OverloadedLists #-} --{-# LANGUAGE NamedFieldPuns #-} module FV.Types ( MCtruth (..) , Prong (..) , Chi2 (..) , VHMeas (..), hFilter , XMeas (..), vBlowup, xXMeas, yXMeas, zXMeas , DMeas (..), Pos, distance , HMeas (..) , QMeas (..), fromHMeas , PMeas (..), fromQMeas, invMass , MMeas (..) , XFit (..) , chi2Vertex, zVertex, rVertex, z0Helix ) where import Prelude.Extended import qualified Data.Vector.Unboxed as A ( drop, zip, map, foldl ) import Data.Foldable ( fold ) import Data.Cov import Data.Cov.Jac ( Jac (..) ) ----------------------------------------------- -- Prong -- a prong results from a vertex fit of N helices newtype Chi2 = Chi2 Number instance Show Chi2 where show (Chi2 c2) = show c2 instance Semiring Chi2 where add (Chi2 c0) (Chi2 c1) = Chi2 (c0+c1) zero = Chi2 0.0 mul (Chi2 c0) (Chi2 c1) = Chi2 (c0*c1) one = Chi2 1.0 instance Num Chi2 where fromInteger i = Chi2 $ fromInteger i negate (Chi2 c) = Chi2 (negate c) abs (Chi2 c) = Chi2 (abs c) signum (Chi2 c) = Chi2 (signum c) (*) = error "cannot multiply Chi2*Chi2 to return a Chi2" (+) (Chi2 a) (Chi2 b) = Chi2 (a+b) data Prong = Prong { nProng :: Int , fitVertex :: XMeas , fitMomenta :: List QMeas , fitChi2s :: List Chi2 , measurements :: VHMeas } -- fitMomenta :: Prong -> List QMeas -- fitMomenta (Prong {fitMomenta= f}) = f instance Show Prong where show _ = "Prong!!!!!!!!" ----------------------------------------------- -- VHMeas -- data VHMeas = VHMeas { vertex :: XMeas , helices :: List HMeas } -- vertex :: VHMeas -> XMeas -- vertex (VHMeas {vertex=v}) = v -- helices :: VHMeas -> List HMeas -- helices (VHMeas {helices=hs}) = hs {-- instance Semigroup VHMeas where --} {-- append (VHMeas v hs) (VHMeas _ hs') = VHMeas v ( hs ++ hs' ) --} {-- mempty = VHMeas (XMeas (Matrix.zero 3 1) (Matrix.zero 3 3)) [] --} instance Show VHMeas where show VHMeas {vertex=v_, helices=hs} = "VHMeas w/ " <> show (length hs) <> " tracks. " <> show v_ vBlowup :: Number -> VHMeas -> VHMeas {-- vBlowup scale vm = over vertexLens (blowup scale) vm where --} vBlowup scale VHMeas {vertex= v_, helices= hs} = VHMeas {vertex= blowup scale v_, helices= hs} where blowup :: Number -> XMeas -> XMeas -- blow up diag of cov matrix blowup s (XMeas v__ cv) = XMeas v__ cv' where cv' = scaleDiag s cv hFilter :: List Int -> VHMeas -> VHMeas hFilter is VHMeas {vertex=v_, helices=hs} = VHMeas {vertex=v_, helices= iflt is hs} hRemove :: Int -> VHMeas -> VHMeas hRemove indx VHMeas {vertex=v_, helices=hs} = VHMeas {vertex=v_, helices=irem indx hs} ----------------------------------------------- -- MCtruth -- newtype MCtruth = MCtruth { pu_zpositions :: List Number } instance Show MCtruth where show MCtruth {pu_zpositions=puz} = "MCtruth w/" <> show (length puz) <> " PU z positions." ----------------------------------------------- -- HMeas -- 5-vector and covariance matrix for helix measurement -- data HMeas = HMeas Vec5 Cov5 Number instance Show HMeas where show (HMeas h ch _) = s' where sh = A.map sqrt $ diag ch hs = toArray h s00 = to5fix x <> " +-" <> to5fix dx where x = uidx hs 0 dx = uidx sh 0 s' = A.foldl f s00 (A.drop 1 $ A.zip hs sh) where f s (x, dx) = s <> to3fix x <> " +-" <> to3fix dx z0Helix :: HMeas -> Number z0Helix (HMeas h _ _) = z0 where v = val h z0 = uidx v 4 ----------------------------------------------- -- QMeas -- 3-vector and covariance matrix for momentum measurement -- mπ :: Number mπ = 0.1395675 data QMeas = QMeas Vec3 Cov3 Number instance Show QMeas where show = showQMeas -- print QMeas as a 4-momentum vector with errors, use pt and pz showQMeas :: QMeas -> String showQMeas (QMeas q cq w2pt) = s' where f :: String -> (Number, Number) -> String f s (x, dx) = s <> to3fix x <> " +-" <> to3fix dx m = mπ wp = w2pt qs :: Array Number qs = toArray q w = uidx qs 0 tl = uidx qs 1 psi0 = uidx qs 2 pt = wp / abs w pz = pt*tl psi = psi0*180.0/pi e = sqrt(pt*pt + pz*pz + m*m) jj :: Jac34 jj = Jac { v = [ -wp/w/w, -wp/w/w*tl, 0.0, -(pz*pz + pt*pt)/w/e , 0.0, wp/w, 0.0, pt*pt*tl/e , 0.0, 0.0, 1.0, 0.0], nr = 3 } cq' = jj .*. cq p' = [pt, pz, psi, e] dp = A.map sqrt $ diag cq' d1 = uidx dp 0 d2 = uidx dp 1 d3 = uidx dp 2 d4 = uidx dp 3 dp' = [d1, d2, d3*180.0/pi, d4] s' = A.foldl f "" ( A.zip p' dp' ) <> " GeV" fromHMeas :: HMeas -> QMeas -- just drop the d0, z0 part... fix!!!! fromHMeas (HMeas h ch w2pt) = QMeas q cq w2pt where q = subm 3 h cq = subm2 3 ch h2p :: HMeas -> PMeas h2p = fromQMeas <<< fromHMeas ----------------------------------------------- -- PMeas -- 4-vector and coavariance matrix for momentum px,py,pz and energy -- data PMeas = PMeas Vec4 Cov4 instance Semigroup PMeas where (<>) (PMeas p1 cp1) (PMeas p2 cp2) = PMeas (p1+p2) (cp1 + cp2) instance Monoid PMeas where mappend (PMeas p1 cp1) (PMeas p2 cp2) = PMeas (p1+p2) (cp1 + cp2) mempty = PMeas (fromArray [0.0,0.0,0.0,0.0]) zero instance Show PMeas where show = showPMeas -- print PMeas as a 4-momentum vector px,py,pz,E with errors showPMeas :: PMeas -> String showPMeas (PMeas p cp) = s' where sp = A.map sqrt $ diag cp f s (x, dx) = s <> to3fix x <> " +-" <> to3fix dx -- \xc2b1 ±±±±± s' = A.foldl f "" (A.zip (toArray p) sp) <> " GeV" invMass :: List PMeas -> MMeas invMass = pmass <<< fold pmass :: PMeas -> MMeas pmass (PMeas p cp) = mm where ps = toArray p px = uidx ps 0 py = uidx ps 1 pz = uidx ps 2 e = uidx ps 3 cps = toArray cp c11 = uidx cps 0 c12 = uidx cps 1 c13 = uidx cps 2 c14 = uidx cps 3 c22 = uidx cps 5 c23 = uidx cps 6 c24 = uidx cps 7 c33 = uidx cps 10 c34 = uidx cps 11 c44 = uidx cps 15 m = sqrt $ max (e*e-px*px-py*py-pz*pz) 0.0 sigm0 = px*c11*px + py*c22*py + pz*c33*pz + e*c44*e + 2.0*(px*(c12*py + c13*pz - c14*e) + py*(c23*pz - c24*e) - pz*c34*e) dm = sqrt ( max sigm0 0.0 ) / m mm = MMeas {m=m, dm=dm} fromQMeas :: QMeas -> PMeas fromQMeas (QMeas q0 cq0 w2pt) = PMeas p0 cp0 where m = mπ q0s = toArray q0 w = uidx q0s 0 tl = uidx q0s 1 psi0 = uidx q0s 2 sph = sin psi0 cph = cos psi0 pt = w2pt / abs w px = pt * cph py = pt * sph pz = pt * tl sqr x= x*x e = sqrt(px*px + py*py + pz*pz + m*m) ps = w2pt / w dpdk = ps*ps/w2pt cq0s = toArray cq0 c11 = uidx cq0s 0 c12 = uidx cq0s 1 c13 = uidx cq0s 2 c22 = uidx cq0s 4 c23 = uidx cq0s 5 c33 = uidx cq0s 8 xy = 2.0*ps*dpdk*cph*sph*c13 sxx = sqr (dpdk*cph) * c11 + sqr (ps*sph) * c33 + xy sxy = cph*sph*(dpdk*dpdk*c11 - ps*ps*c33) + ps*dpdk*(sph*sph-cph*cph)*c13 syy = sqr (dpdk*sph) * c11 + sqr (ps*cph) * c33 - xy sxz = dpdk*dpdk*cph*tl*c11 - ps*dpdk*(cph*c12-sph*tl*c13) - ps*ps*sph*c23 syz = dpdk*dpdk*sph*tl*c11 - ps*dpdk*(sph*c12 + cph*tl*c13) + ps*ps*cph*c23 szz = sqr (dpdk*tl) * c11 + ps*ps*c22 - 2.0*ps*dpdk*tl*c12 sxe = (px*sxx + py*sxy + pz*sxz)/e sye = (px*sxy + py*syy + pz*syz)/e sze = (px*sxz + py*syz + pz*szz)/e see = (px*px*sxx + py*py*syy + pz*pz*szz + 2.0*(px*(py*sxy + pz*sxz) + py*pz*syz))/e/e cp0 = fromArray [ sxx, sxy, sxz, sxe , syy, syz, sye , szz, sze , see] p0 = fromArray [px,py,pz,e] ----------------------------------------------- -- MMeas -- scalar mass and error calculated from PMeas covariance matrices -- data MMeas = MMeas { m :: Number , dm :: Number } instance Show MMeas where show MMeas {m=m, dm=dm} = " " <> to1fix (m*1000.0) <> " +-" <> to1fix (dm*1000.0) <> " MeV" ----------------------------------------------- -- XMeas -- 3-vector and covariance matrix for position/vertex measurement -- data DMeas = DMeas Number Number -- distance and error instance Show DMeas where show (DMeas d sd) = to2fix d <> " +-" <> to2fix sd <> " cm" class Pos a where distance :: a -> a -> DMeas -- distance between two positions instance Pos XMeas where distance (XMeas v0 vv0) (XMeas v1 vv1) = DMeas d sd where v0s = toArray v0 x0 = uidx v0s 0 y0 = uidx v0s 1 z0 = uidx v0s 2 v1s = toArray v1 x1 = uidx v1s 0 y1 = uidx v1s 1 z1 = uidx v1s 2 d = sqrt(sqr(x0-x1) + sqr(y0-y1) + sqr(z0-z1)) dd :: Vec3 dd = fromArray [(x0-x1)/d, (y0-y1)/d, (z0-z1)/d] tem0 = dd .*. vv0 tem1 = dd .*. vv1 sd = sqrt (tem0 + tem1) data XMeas = XMeas Vec3 Cov3 instance Semigroup XMeas where (<>) (XMeas x1 cx1) (XMeas x2 cx2) = XMeas (x1 + x2) (cx1 + cx2) instance Monoid XMeas where mempty = XMeas zero zero mappend = undefined instance Show XMeas where show = showXMeas -- return a string showing vertex position vector with errors showXMeas :: XMeas -> String showXMeas (XMeas v cv) = s' where vv = toArray v x = uidx vv 0 y = uidx vv 1 z = uidx vv 2 s2v = A.map sqrt $ diag cv dx = uidx s2v 0 dy = uidx s2v 1 dz = uidx s2v 2 f :: Number -> Number -> String -> String f x dx s = s <> to2fix x <> " +-" <> to2fix dx s' = f z dz <<< f y dy <<< f x dx $ "(r,z) =" <> "(" <> to2fix (sqrt (x*x + y*y)) <> ", " <> to2fix z <> "), x y z =" xXMeas :: XMeas -> Number xXMeas (XMeas v _) = x where x = uidx (val v) 0 yXMeas :: XMeas -> Number yXMeas (XMeas v _) = y where y = uidx (val v) 1 zXMeas :: XMeas -> Number zXMeas (XMeas v _) = z where z = uidx (val v) 2 data XFit = XFit Vec3 Cov3 Chi2 instance Show XFit where show (XFit x xx c2) = showXMeas (XMeas x xx) <> ", chi2=" <> show c2 chi2Vertex :: XFit -> Chi2 chi2Vertex (XFit _ _ c2) = c2 zVertex :: XFit -> Double zVertex (XFit v _ _) = z where z = uidx (val v) 2 rVertex :: XFit -> Double rVertex (XFit v _ _) = r where r = sqrt $ (sqr (uidx (val v) 0)) + (sqr (uidx (val v) 1))
LATBauerdick/fv.hs
src/FV/Types.hs
bsd-3-clause
10,884
1
20
3,200
4,393
2,343
2,050
277
1
-- !!! local aliases module M where import qualified Data.OldList as M import qualified Data.Maybe as M x = M.length b = M.isJust
jstolarek/ghc
testsuite/tests/module/mod106.hs
bsd-3-clause
132
0
5
25
35
24
11
5
1
module Main where main :: IO() main = do putStrLn "Give me a list, and I'll give you it's last...Maybe" ls <- getLine print $ myLast ls myLast :: [a] -> Maybe a myLast [] = Nothing myLast [x] = Just x myLast (_:xs) = myLast xs myLast' :: [a] -> a myLast' [] = error "Cannot operate on empty lists, exiting..." myLast' [x] = x myLast' (_:xs) = myLast' xs
Jaso-N7/H99-solutions
1-10/1.hs
bsd-3-clause
375
0
8
92
153
77
76
14
1
{-# LANGUAGE CPP, FlexibleInstances, RankNTypes, TypeSynonymInstances #-} {-# OPTIONS -Wall #-} -- | An extension module of building blocks. Contains booleans, comparison operations, branchings. module Language.Paraiso.OM.Builder.Boolean (eq, ne, lt, le, gt, ge, select) where import Data.Dynamic (Typeable, typeOf) import qualified Language.Paraiso.OM.Arithmetic as A import Language.Paraiso.OM.Builder.Internal import Language.Paraiso.OM.DynValue as DVal import Language.Paraiso.OM.Graph import Language.Paraiso.OM.Realm as Realm import Language.Paraiso.OM.Value as Val infix 4 `eq`, `ne`, `lt`, `le`, `gt`, `ge` -- | generate a binary operator that returns Bool results. mkOp2B :: (TRealm r, Typeable c) => A.Operator -- ^The operation to be performed -> (Builder v g a (Value r c)) -- ^The first argument -> (Builder v g a (Value r c)) -- ^The second argument -> (Builder v g a (Value r Bool)) -- ^The result mkOp2B op builder1 builder2 = do v1 <- builder1 v2 <- builder2 let r1 = Val.realm v1 n1 <- valueToNode v1 n2 <- valueToNode v2 n0 <- addNodeE [n1, n2] $ NInst (Arith op) n01 <- addNodeE [n0] $ NValue (toDyn v1){typeRep = typeOf True} return $ FromNode r1 True n01 type CompareOp = (TRealm r, Typeable c) => (Builder v g a (Value r c)) -> (Builder v g a (Value r c)) -> (Builder v g a (Value r Bool)) -- | Equal eq :: CompareOp eq = mkOp2B A.EQ -- | Not equal ne :: CompareOp ne = mkOp2B A.NE -- | Less than lt :: CompareOp lt = mkOp2B A.LT -- | Less than or equal to le :: CompareOp le = mkOp2B A.LE -- | Greater than gt :: CompareOp gt = mkOp2B A.GT -- | Greater than or equal to ge :: CompareOp ge = mkOp2B A.GE -- | selects either the second or the third argument based select ::(TRealm r, Typeable c) => (Builder v g a (Value r Bool)) -- ^The 'Bool' condition -> (Builder v g a (Value r c)) -- ^The value chosen when the condition is 'True' -> (Builder v g a (Value r c)) -- ^The value chosen when the condition is 'False' -> (Builder v g a (Value r c)) -- ^The result select builderB builder1 builder2 = do vb <- builderB v1 <- builder1 v2 <- builder2 nb <- valueToNode vb n1 <- valueToNode v1 n2 <- valueToNode v2 n0 <- addNodeE [nb, n1, n2] $ NInst (Arith A.Select) n01 <- addNodeE [n0] $ NValue (toDyn v1) let r1 = Val.realm v1 c1 = Val.content v1 return $ FromNode r1 c1 n01
drmaruyama/Paraiso
Language/Paraiso/OM/Builder/Boolean.hs
bsd-3-clause
2,474
0
12
589
834
450
384
60
1
{-# LANGUAGE FlexibleContexts #-} {-# LANGUAGE TypeApplications #-} import Control.Monad.Classes hiding (exec) import Control.Monad.Classes.Log import Control.Monad.Log (LoggingT, runLoggingT) import Control.Monad.Trans.Reader (ReaderT(..)) import Control.Monad.Trans.State (State, execState) import Test.Tasty import Test.Tasty.HUnit main :: IO () main = defaultMain $ testGroup "tests" [testCase "should compile" testcase ,testCase "writer impl" writecase] where testcase = assertEqual "correct results" opresults (exec op) writecase = assertEqual "correct writes" opresults (exec opw) opresults = [RRes "bar", LRes "foo"] data Foo = Foo String data Bar = Bar String data Res = LRes String | RRes String deriving (Eq, Ord, Show) logit :: (MonadReader a m, MonadLog a m) => (a -> a) -> m () logit f = ask >>= logMessage . f -- | This is the operation that will not compile under mtl -- | due to functional dependencies. op :: (MonadReader Foo m, MonadReader Bar m, MonadLog Foo m, MonadLog Bar m) => m () op = logit (id @Foo) >> logit (id @Bar) -- | handler for Foo foo :: MonadState [Res] m => Foo -> m () foo (Foo s) = modify $ \r -> LRes s : r -- | Handler for Bar bar :: MonadState [Res] m => Bar -> m () bar (Bar s) = modify $ \r -> RRes s : r exec :: ReaderT Foo (ReaderT Bar (LoggingT Foo (LoggingT Bar (State [Res])))) () -> [Res] exec m = execState (runLoggingT (runLoggingT (runReaderT (runReaderT m (Foo "foo")) (Bar "bar")) foo) bar) [] writeit :: (MonadReader a m, MonadWriter a m) => (a -> a) -> m () writeit f = ask >>= tell . f opw :: (MonadReader Foo m, MonadReader Bar m, MonadWriter Foo m, MonadWriter Bar m) => m () opw = writeit (id @Foo) >> writeit (id @Bar)
edwardgeorge/monad-classes-logging
test/Spec.hs
bsd-3-clause
1,754
0
15
365
713
377
336
33
1
{-# OPTIONS -Wall #-} module Main where import Codec.Picture( PixelRGBA8( .. ), writePng ) import Graphics.Rasterific import Graphics.Rasterific.Texture import qualified Data.Vector as V import qualified Data.Vector.Unboxed as UV import qualified Data.Vector.Unboxed.Mutable as UM import Control.Monad.Primitive import Control.Monad (filterM) import Data.Word (Word8) import System.Random type Width = Int type Height = Int type XPos = Int type YPos = Int type Color = (Word8, Word8, Word8) type P2 = (XPos, YPos) type Pixel = (P2, Color, Bool) type Canvas m = UM.MVector m Pixel type Points = V.Vector P2 width :: Width width = 640 height :: Height height = 360 generatePixels :: Width -> Height -> [Pixel] generatePixels w h = [((j,i),(255,255,255),False) | i <- [0..h-1], j <- [0..w-1]] createPixelInstruction :: Pixel -> Drawing PixelRGBA8 () createPixelInstruction ((x,y),(r,g,b),_) = withTexture (uniformTexture color) . fill $ rectangle point 1 1 where color = PixelRGBA8 (r :: Word8) (g :: Word8) (b :: Word8) 255 point = V2 (fromIntegral x) (fromIntegral y) generateCanvas :: PrimMonad m => Width -> Height -> StdGen -> m (Canvas (PrimState m)) generateCanvas w h g = do let seed = (quot w 2, quot h 2) seedColor = (207,0,15) canvasPoints = V.fromList [(j,i) | i <- [0..height-1], j <- [0..width-1]] mV <- UV.thaw $ UV.fromList $ generatePixels w h -- create canvas _ <- drawPixel mV seedColor seed -- draw center pixel (seed) _ <- draw mV canvasPoints seedColor g return mV draw :: PrimMonad m => Canvas (PrimState m) -> Points -> Color -> StdGen -> m (Canvas (PrimState m)) draw canvas canvasPoints seedColor g = do let (nextColor,g') = newColor seedColor g points <- touchedPoints canvas canvasPoints V.sequence_ $ V.map (drawBorderPixels canvas nextColor g') points if (V.length points < (width*height)) then do (draw canvas canvasPoints nextColor g') else return canvas randVals :: StdGen -> ((Double, Double, Double), StdGen) randVals g = ((x,y,z),g''') where (x, g' ) = randomR (0,1) g (y, g'' ) = randomR (0,1) g' (z, g''') = randomR (0,1) g'' newColor :: Color -> StdGen -> (Color,StdGen) newColor c g = (calculateColor c prob, g') where (prob,g') = randVals g calculateColor :: Color -> (Double, Double, Double) -> Color calculateColor (r,g,b) (x,y,z) = color where color = (incrementColor r (inc x), incrementColor g (inc y), incrementColor b (inc z)) incrementColor :: Word8 -> Inc -> Word8 incrementColor w i | i == Add && w == 255 = 244 | i == Subtract && w == 0 = 1 | i == Add = w + 5 | i == Subtract = w - 5 | otherwise = w data Inc = Add | Same | Subtract deriving (Eq) inc :: Double -> Inc inc d | d >= 0.0 && d < 0.33 = Subtract | d >= 0.33 && d < 0.66 = Same | otherwise = Add drawBorderPixels :: PrimMonad m => Canvas (PrimState m) -> Color -> StdGen -> P2 -> m (Canvas (PrimState m)) drawBorderPixels canvas color g point = do (_,_,touched) <- UM.read canvas (index point) if touched then do points <- pointsToDraw canvas point g sequence_ $ (drawPixel canvas color) <$> points return canvas else return canvas -- Generate list of safe pixels to check pointsToDraw :: PrimMonad m => Canvas (PrimState m) -> P2 -> StdGen -> m [P2] pointsToDraw canvas (x,y) g = do let validPoints = filter validPoint points availablePoints <- untouchedPoints canvas validPoints let shuffledPoints = shuffle availablePoints g takenPoints = take 1 shuffledPoints return takenPoints where n = (x,y-1) ne = (x+1,y-1) e = (x+1,y) se = (x+1,y+1) s = (x,y+1) sw = (x-1,y+1) w = (x-1,y) nw = (x-1,y-1) points = [n,ne,e,se,s,sw,w,nw] shuffle :: [P2] -> StdGen -> [P2] shuffle [] _ = [] shuffle xs g = val : shuffle rest g' where (idx,g') = randomR (0,length xs - 1) g val = xs !! idx bef = take idx xs aft = drop (idx+1) xs rest = bef ++ aft validPoint :: P2 -> Bool validPoint (x,y) | x < 0 || x >= width = False | y < 0 || y >= height = False | otherwise = True touchedPoints :: PrimMonad m => Canvas (PrimState m) -> Points -> m Points touchedPoints canvas points = do validPoints <- V.filterM (touchedPoint canvas) points return validPoints --getTouchedPoint :: PrimMonad m => Canvas (PrimState m) -> Points -> m P2 --getTouchedPoint canvas points = do -- validPoints <- V.filterM (touchedPoint canvas) points -- return $ V.head $ V.take 1 validPoints --touchedPoint :: PrimMonad m => Canvas (PrimState m) -> P2 -> m Bool --touchedPoint canvas point = do -- (_,_,touched) <- UM.read canvas (index point) -- if touched then return True else return False touchedPoint :: PrimMonad m => Canvas (PrimState m) -> P2 -> m Bool touchedPoint canvas n = do (_,_,touched) <- UM.read canvas $ index n if touched then return True else return False untouchedPoints :: PrimMonad m => Canvas (PrimState m) -> [P2] -> m [P2] untouchedPoints canvas points = do validPoints <- filterM (untouchedPoint canvas) points return validPoints untouchedPoint :: PrimMonad m => Canvas (PrimState m) -> P2 -> m Bool untouchedPoint canvas point = do (_,_,touched) <- UM.read canvas (index point) if touched then return False else return True drawPixel :: PrimMonad m => Canvas (PrimState m) -> Color -> P2 -> m (Canvas (PrimState m)) drawPixel canvas color point = do let pixel = (point, color, True) UM.write canvas (index point) pixel return canvas vecToInstructions :: UV.Vector Pixel -> [Drawing PixelRGBA8 ()] vecToInstructions v = createPixelInstruction <$> UV.toList v index :: P2 -> Int index (x,y) = width * y + x main :: IO () main = do gen <- getStdGen mutableCanvas <- generateCanvas width height gen frozenCanvas <- UV.freeze mutableCanvas let drawInstructions = vecToInstructions frozenCanvas white = PixelRGBA8 255 255 255 255 img = renderDrawing width height white $ sequence_ drawInstructions writePng "image.png" img
ismailmustafa/Paint
src/Main.hs
bsd-3-clause
6,243
0
15
1,507
2,486
1,304
1,182
142
2
{-# LANGUAGE FlexibleInstances #-} module Shader where import Prelude hiding (any, floor, ceiling) import Control.Monad (when) import Control.Monad.IO.Class (MonadIO, liftIO) import Foreign import Foreign.C.String import Graphics.GL import Linear import Paths_hadoom import Util import qualified Data.ByteString as BS import qualified Data.Text.Encoding as Text import qualified Data.Text.IO as Text newtype GLProgram = GLProgram {unGLProgram :: GLuint} createShaderProgram :: FilePath -> FilePath -> IO GLProgram createShaderProgram vertexShaderPath fragmentShaderPath = do vertexShader <- glCreateShader GL_VERTEX_SHADER compileShader vertexShaderPath vertexShader fragmentShader <- glCreateShader GL_FRAGMENT_SHADER compileShader fragmentShaderPath fragmentShader shaderProg <- glCreateProgram glAttachShader shaderProg vertexShader glAttachShader shaderProg fragmentShader mapM_ (\(x,y) -> withCString x (glBindAttribLocation shaderProg y)) [("in_Position",positionAttribute) ,("in_Normal",normalAttribute) ,("in_Tangent",tangentAttribute) ,("in_Bitangent",bitangentAttribute) ,("in_UV",uvAttribute)] glLinkProgram shaderProg linked <- overPtr (glGetProgramiv shaderProg GL_LINK_STATUS) when (linked == GL_FALSE) (do maxLength <- overPtr (glGetProgramiv shaderProg GL_INFO_LOG_LENGTH) logLines <- allocaArray (fromIntegral maxLength) (\p -> alloca (\lenP -> do glGetProgramInfoLog shaderProg maxLength lenP p len <- peek lenP peekCStringLen (p,fromIntegral len))) putStrLn logLines) return (GLProgram shaderProg) where compileShader path shader = do src <- getDataFileName path >>= Text.readFile BS.useAsCString (Text.encodeUtf8 src) (\ptr -> withArray [ptr] (\srcs -> glShaderSource shader 1 srcs nullPtr)) glCompileShader shader -- GL.get (GL.shaderInfoLog shader) >>= putStrLn positionAttribute, uvAttribute, normalAttribute, tangentAttribute, bitangentAttribute :: GLuint positionAttribute = 0 normalAttribute = 1 tangentAttribute = 2 bitangentAttribute = 3 uvAttribute = 4 class Uniform a where setUniformLoc :: MonadIO m => GLint -> a -> m () setUniform :: (MonadIO m, Uniform a) => GLProgram -> String -> a -> m () setUniform (GLProgram shaderProg) name a = do loc <- liftIO (withCString name (glGetUniformLocation shaderProg)) glUseProgram shaderProg setUniformLoc loc a instance Storable a => Uniform (M44 a) where setUniformLoc loc m = liftIO (with m (glUniformMatrix4fv loc 1 0 . castPtr)) instance Uniform (V2 GLfloat) where setUniformLoc loc (V2 x y) = glUniform2f loc x y instance Uniform (V4 GLfloat) where setUniformLoc loc (V4 x y z w) = glUniform4f loc x y z w instance Uniform GLint where setUniformLoc = glUniform1i instance Uniform GLfloat where setUniformLoc = glUniform1f getSubroutineIndex :: MonadIO m => GLProgram -> String -> m GLuint getSubroutineIndex (GLProgram p) subroutine = liftIO (withCString subroutine (glGetSubroutineIndex p GL_FRAGMENT_SHADER)) getUniformBlockIndex :: MonadIO m => GLProgram -> String -> m GLuint getUniformBlockIndex (GLProgram p) block = liftIO (withCString block (glGetUniformBlockIndex p))
ocharles/hadoom
hadoom/Shader.hs
bsd-3-clause
3,772
0
22
1,085
957
493
464
92
1
{-# LANGUAGE RecordWildCards #-} module TemplateParse (parseChordTemplate) where import Text.ParserCombinators.Parsec import Control.Applicative ((<$>), (*>), (<*), (<$), (<*>)) import Data.Maybe (fromMaybe, isJust, catMaybes) import Note (ABC (..), Octave (..)) import Interval (Interval (..)) import qualified Interval as In import Pattern -- <chord> :: <root_note>["("<octave_range>")"](<explicit_intervals>|<chord_spec>) -- <octave_range> :: \d|(\d,\d,...)|(\d-\d) -- <explicit_intervals> :: "{" (<interval_spec> ",")* "}" -- <interval_spec> :: -- {0=,4,7?,11=?} -- 0 exact, 4 with inversions allowed, optional 7 with inversions, optional 11 exact -- {0=,4,*}-{10,11} -- 0, 4 and any other, except 10 and 11 -- {0=,4,7?,10|11} -- ..., 10 or 11 -- {} -- <chord_spec> :: <chord_quality><interval_num><altered_fifth><additional_interval> -- <chord_quality> :: "m" | "M" | "" -- <interval_num> :: "(" ("maj" | "min" | "aug" | "dim")<number> ")" infix 0 <??> (<??>) = flip (<?>) oneOfStr :: [String] -> CharParser () String oneOfStr ss = choice (map string ss) pChord :: CharParser () (ChordPattern NotePattern [IntervalPattern]) pChord = do note <- "note pattern" <??> pTemplateValue pNote octave <- "octave pattern" <??> fromMaybe Any <$> optionMaybe (char '-' *> pTemplateValue pOctave) intervalPatterns <- "chord spec" <??> pExplicitIntervals <|> (chordSpecToIntervals <$> pChordSpec) return $ ChordPattern (NotePattern note octave) intervalPatterns pNote :: CharParser () ABC pNote = do n <- oneOf "ABCDEFG" m <- maybe [] ((: []) . sharpOrFlat) <$> optionMaybe (oneOf "#b") return $ read (n : m) where sharpOrFlat '#' = 's' sharpOrFlat x = x pOctave :: CharParser () Octave pOctave = Octave . read . (: []) <$> oneOf "012345678" pIntervalName :: CharParser () Interval pIntervalName = "interval name" <??> choiceInterval [(In.perf1, ["P1", "perf1", "dim2"]) ,(In.min2, ["m2", "min2", "A1", "aug1"]) ,(In.maj2, ["M2", "maj2", "d3", "dim3"]) ,(In.min3, ["m3", "min3", "A2", "aug2"]) ,(In.maj3, ["M3", "maj3", "d4", "dim4"]) ,(In.perf4, ["P4", "perf4", "A3", "aug3"]) ,(In.dim5, ["d5", "dim5", "A4", "aug4", "tritone"]) ,(In.perf5, ["P5", "perf5", "d6", "dim6"]) ,(In.min6, ["m6", "min6", "A5", "aug5"]) ,(In.maj6, ["M6", "maj6", "d7", "dim7"]) ,(In.min7, ["m7", "min7", "A6", "aug6"]) ,(In.maj7, ["M7", "maj7", "d8", "dim8"]) ,(In.perf8, ["P8", "perf8", "A7", "aug7"]) ,(In.Interval 13, ["m9", "min9"]) -- p8 + min2 ,(In.Interval 14, ["M9", "maj9"]) -- p8 + maj2 ,(In.Interval 15, ["m10", "min10"]) -- p8 + min3 ,(In.Interval 16, ["M10", "maj10"]) -- p8 + maj3 ] where p_oneName (interval, names) = try $ interval <$ oneOfStr names choiceInterval is = choice (map p_oneName is) pIntervalInteger :: CharParser () Interval pIntervalInteger = (Interval . read) <$> many1 digit pInterval :: CharParser () Interval pInterval = pIntervalInteger <|> pIntervalName pIntervalPattern :: CharParser () IntervalPattern pIntervalPattern = do noInversions <- isJust <$> optionMaybe (char '=') intervalDef <- pInterval isOptional <- isJust <$> optionMaybe (char '?') let ipv = IntervalPatternValue { interval = intervalDef, inversionsAllowed = not noInversions } let po = if isOptional then OneOrNone ipv else ExactlyOne ipv return $ po pExplicitIntervals :: CharParser () [IntervalPattern] pExplicitIntervals = char '{' *> sepBy1 pIntervalPattern (char ',') <* char '}' pTemplateValue :: (Bounded a, Enum a) => CharParser () a -> CharParser () (PatternValue a) pTemplateValue p = Any <$ char '*' <|> maybe Any OneOf <$> optionMaybe oneOrManyV where oneV = (: []) <$> p manyV = "pattern" <??> char '[' *> sepBy1 p (char ',') <* char ']' oneOrManyV = manyV <|> oneV data ChordQuality = MinorChord | MajorChord data IntervalQuality = Minor | Major | Diminished | Augmented | Dominant data ChordSpec = ChordSpec { chordQuality :: Maybe ChordQuality, mainInterval :: Maybe (Maybe IntervalQuality, Integer), addInterval :: Maybe Integer } pChordSpec :: CharParser () ChordSpec pChordSpec = createSpec <$> pChordQuality <*> pIntervalSpec <*> pAdditionalInterval where createSpec q mi ai = ChordSpec { chordQuality = q, mainInterval = mi, addInterval = ai } pChordQuality = "chord quality" <??> optionMaybe ((try $ MinorChord <$ char 'm') <|> (try $ MajorChord <$ char 'M')) pIntervalQuality = (<??>) "main chord interval" $ -- @TODO seems broken optionMaybe $ choice (map try [Minor <$ (string "min") ,Major <$ (string "maj") ,Diminished <$ (string "dim") ,Augmented <$ (string "aug") ,Dominant <$ (string "dom")]) pIntervalNumber = read <$> many1 digit pAdditionalInterval = (<??>) "additional chord note" $ optionMaybe $ read <$> (string "add" *> many1 digit) pIntervalSpec = optionMaybe $ (,) <$> (char '(' *> pIntervalQuality) <*> (pIntervalNumber <* char ')') chordSpecToIntervals :: ChordSpec -> [IntervalPattern] chordSpecToIntervals (ChordSpec {..}) = root : (catMaybes $ third : fifth : main : added : []) where ipat x y = IntervalPatternValue { interval = x, inversionsAllowed = y } inversions x = ipat x True noInversions x = ipat x False root = ExactlyOne $ noInversions (In.Interval 0) chordQ = fromMaybe MajorChord chordQuality -- @TODO ugly stuff makeInterval (q, n) = In.Interval $ neutralI + modI where neutralI = case n of 7 -> 11 6 -> 9 9 -> 14 10 -> 16 11 -> 17 -- ?? modI = case fromMaybe Major q of Major -> 0 Minor -> -1 Augmented -> 1 Diminished -> -2 Dominant -> error "ugh" third = Just . ExactlyOne . noInversions $ (case chordQ of MajorChord -> In.maj3 MinorChord -> In.min3) fifth = Just . ExactlyOne . inversions $ In.perf5 main = ExactlyOne . inversions . makeInterval <$> mainInterval added = Nothing parseChordTemplate :: String -> Either ParseError (ChordPattern NotePattern [IntervalPattern]) parseChordTemplate = parse pChord "(chord def)"
blacktaxi/inversion
src/TemplateParse.hs
bsd-3-clause
6,807
0
14
1,900
1,973
1,089
884
137
10
{-# LANGUAGE ExtendedDefaultRules #-} {-# LANGUAGE PolyKinds #-} {-# LANGUAGE ConstraintKinds #-} {-# LANGUAGE UndecidableInstances #-} {-# OPTIONS_GHC -Wall #-} -- | Algebra module Tower.Ring ( -- * Ring Semiring , Ring , CRing ) where import Protolude (Double, Float, Int, Integer,Bool(..)) import Data.Functor.Rep import Tower.Additive import Tower.Multiplicative import Tower.Distribution -- | a semiring class ( Additive a , MultiplicativeAssociative a , MultiplicativeUnital a , Distribution a) => Semiring a instance Semiring Double instance Semiring Float instance Semiring Int instance Semiring Integer instance Semiring Bool instance (Representable r, Semiring a) => Semiring (r a) -- | Ring class ( AdditiveGroup a , MultiplicativeAssociative a , MultiplicativeUnital a , Distribution a) => Ring a instance Ring Double instance Ring Float instance Ring Int instance Ring Integer instance (Representable r, Ring a) => Ring (r a) -- | CRing is a Commutative Ring. It arises often due to * being defined as only multiplicative commutative, yet fromInteger being a `Integer -> Ring` (and thus not necessarily commutative). class ( Multiplicative a, Ring a) => CRing a instance CRing Double instance CRing Float instance CRing Int instance CRing Integer instance (Representable r, CRing a) => CRing (r a)
tonyday567/tower
src/Tower/Ring.hs
bsd-3-clause
1,384
0
7
268
337
176
161
-1
-1
module Main where import GameIO import System.Console.ANSI main :: IO () main = runGame
MikePors/FizzBuzzWarz
app/Main.hs
bsd-3-clause
90
0
6
16
28
17
11
5
1
{-# OPTIONS_GHC -Wall #-} module Canonicalize.Variable where import qualified Data.Either as Either import qualified Data.Map as Map import qualified Data.Set as Set import qualified AST.Helpers as Help import qualified AST.Module.Name as ModuleName import qualified AST.Type as Type import qualified AST.Variable as Var import qualified Reporting.Annotation as A import qualified Reporting.Error.Canonicalize as Error import qualified Reporting.Error.Helpers as Error import qualified Reporting.Region as R import qualified Canonicalize.Environment as Env import qualified Canonicalize.Result as Result import Elm.Utils ((|>)) variable :: R.Region -> Env.Environment -> String -> Result.ResultErr Var.Canonical variable region env var = case toVarName var of Right (name, varName) | ModuleName.isNative name -> error "TODO" "Result.var (Var.Canonical (Var.Module name) varName)" _ -> case Set.toList `fmap` Map.lookup var (Env._values env) of Just [v] -> Result.var v Just vs -> preferLocals region env "variable" vs var Nothing -> notFound region "variable" (Map.keys (Env._values env)) var tvar :: R.Region -> Env.Environment -> String -> Result.ResultErr (Either Var.Canonical (Var.Canonical, [String], Type.Canonical) ) tvar region env var = case adts ++ aliases of [] -> notFound region "type" (Map.keys (Env._adts env) ++ Map.keys (Env._aliases env)) var [v] -> Result.var' extract v vs -> preferLocals' region env extract "type" vs var where adts = map Left (maybe [] Set.toList (Map.lookup var (Env._adts env))) aliases = map Right (maybe [] Set.toList (Map.lookup var (Env._aliases env))) extract value = case value of Left v -> v Right (v,_,_) -> v pvar :: R.Region -> Env.Environment -> String -> Int -> Result.ResultErr Var.Canonical pvar region env var actualArgs = case Set.toList `fmap` Map.lookup var (Env._patterns env) of Just [value] -> foundArgCheck value Just values -> preferLocals' region env fst "pattern" values var `Result.andThen` foundArgCheck Nothing -> notFound region "pattern" (Map.keys (Env._patterns env)) var where foundArgCheck (name, expectedArgs) = if actualArgs == expectedArgs then Result.var name else Result.err (A.A region (Error.argMismatch name expectedArgs actualArgs)) -- FOUND preferLocals :: R.Region -> Env.Environment -> String -> [Var.Canonical] -> String -> Result.ResultErr Var.Canonical preferLocals region env = preferLocals' region env id preferLocals' :: R.Region -> Env.Environment -> (a -> Var.Canonical) -> String -> [a] -> String -> Result.ResultErr a preferLocals' region env extract kind possibilities var = case filter (isLocal (Env._home env) . extract) possibilities of [] -> ambiguous possibilities [v] -> Result.var' extract v --Local vars shadow top-level vars [v1, v2] | isTopLevel (extract v1) -> Result.var' extract v2 | isTopLevel (extract v2) -> Result.var' extract v1 locals -> ambiguous locals where ambiguous possibleVars = Result.err (A.A region (Error.variable kind var Error.Ambiguous vars)) where vars = map (Var.toString . extract) possibleVars isLocal :: ModuleName.Canonical -> Var.Canonical -> Bool isLocal contextName (Var.Canonical home _) = case home of Var.Local -> True Var.TopLevel _ -> True Var.BuiltIn -> False Var.Module name -> name == contextName isTopLevel :: Var.Canonical -> Bool isTopLevel (Var.Canonical home _) = case home of Var.TopLevel _ -> True _ -> False -- NOT FOUND HELPERS type VarName = Either String (ModuleName.Raw, String) toVarName :: String -> VarName toVarName var = case Help.splitDots var of [x] -> Left x xs -> Right (init xs, last xs) noQualifier :: VarName -> String noQualifier name = case name of Left x -> x Right (_, x) -> x qualifiedToString :: (ModuleName.Raw, String) -> String qualifiedToString (modul, name) = ModuleName.toString modul ++ "." ++ name isOp :: VarName -> Bool isOp name = Help.isOp (noQualifier name) -- NOT FOUND notFound :: R.Region -> String -> [String] -> String -> Result.ResultErr a notFound region kind possibilities var = let name = toVarName var possibleNames = map toVarName possibilities (problem, suggestions) = case name of Left _ -> exposedProblem name possibleNames Right (modul, varName) -> qualifiedProblem modul varName (Either.rights possibleNames) in Result.err (A.A region (Error.variable kind var problem suggestions)) exposedProblem :: VarName -> [VarName] -> (Error.VarProblem, [String]) exposedProblem name possibleNames = let (exposed, qualified) = possibleNames |> filter (\n -> isOp name == isOp n) |> Error.nearbyNames noQualifier name |> Either.partitionEithers in ( Error.ExposedUnknown , exposed ++ map qualifiedToString qualified ) qualifiedProblem :: ModuleName.Raw -> String -> [(ModuleName.Raw, String)] -> (Error.VarProblem, [String]) qualifiedProblem moduleName name allQualified = let availableModules = Set.fromList (map fst allQualified) moduleNameString = ModuleName.toString moduleName in case Set.member moduleName availableModules of True -> ( Error.QualifiedUnknown moduleNameString name , allQualified |> filter ((==) moduleName . fst) |> map snd |> Error.nearbyNames id name ) False -> ( Error.UnknownQualifier moduleNameString name , Set.toList availableModules |> map ModuleName.toString |> Error.nearbyNames id moduleNameString )
pairyo/elm-compiler
src/Canonicalize/Variable.hs
bsd-3-clause
6,338
0
21
1,834
1,911
987
924
183
4
module Main where main :: IO () main = putStrLn "sup"
bts/free-transformers
app/Main.hs
bsd-3-clause
55
0
6
12
22
12
10
3
1
{-# LANGUAGE OverloadedStrings, CPP #-} module HPACK.HuffmanSpec where #if __GLASGOW_HASKELL__ < 709 import Control.Applicative ((<$>)) #endif import Data.ByteString (ByteString) import qualified Data.ByteString.Base16 as B16 import qualified Data.ByteString.Char8 as BS import Data.Char (toLower) import Network.HPACK import Network.HPACK.Internal import Test.Hspec import Test.Hspec.QuickCheck testData :: [(ByteString, ByteString)] testData = [ ("", "") , ("www.example.com", "f1e3c2e5f23a6ba0ab90f4ff") , ("no-cache", "a8eb10649cbf") , ("custom-key", "25a849e95ba97d7f") , ("custom-value", "25a849e95bb8e8b4bf") , ("private", "aec3771a4b") , ("Mon, 21 Oct 2013 20:13:21 GMT", "d07abe941054d444a8200595040b8166e082a62d1bff") , ("https://www.example.com", "9d29ad171863c78f0b97c8e9ae82ae43d3") , ("Mon, 21 Oct 2013 20:13:22 GMT", "d07abe941054d444a8200595040b8166e084a62d1bff") , ("gzip", "9bd9ab") , ("foo=ASDJKHQKBZXOQWEOPIUAXQWEOIU; max-age=3600; version=1", "94e7821dd7f2e6c7b335dfdfcd5b3960d5af27087f3672c1ab270fb5291f9587316065c003ed4ee5b1063d5007") ] shouldBeEncoded :: ByteString -> ByteString -> Expectation shouldBeEncoded inp out = do out' <- BS.map toLower . B16.encode <$> encodeHuffman inp out' `shouldBe` out shouldBeDecoded :: ByteString -> ByteString -> Expectation shouldBeDecoded inp out = do out' <- decodeHuffman $ B16.decodeLenient inp out' `shouldBe` out tryDecode :: ByteString -> IO ByteString tryDecode inp = decodeHuffman $ B16.decodeLenient inp spec :: Spec spec = do describe "encode and decode" $ do prop "duality" $ \cs -> do let bs = BS.pack cs es <- encodeHuffman bs ds <- decodeHuffman es ds `shouldBe` bs describe "encode" $ do it "encodes" $ do mapM_ (\(x,y) -> x `shouldBeEncoded` y) testData describe "decode" $ do it "decodes" $ do tryDecode "ff" `shouldThrow` (== TooLongEos) tryDecode "ffffeaff" `shouldThrow` (== TooLongEos) "ffffea" `shouldBeDecoded` "\9" mapM_ (\(x,y) -> y `shouldBeDecoded` x) testData
kazu-yamamoto/http2
test/HPACK/HuffmanSpec.hs
bsd-3-clause
2,149
0
18
426
548
306
242
51
1
module Graphics.Hexif.Utils where import Data.Binary import Data.Binary.Get import qualified Data.ByteString.Lazy as BL import Data.Hex -- | The encoding of the binary data. -- Motorola is big endian, Intel is low endian data Encoding = Intel | Motorola deriving (Show) -- | Little support function to read 16 bit integers getWord16 :: Encoding -> Get Word16 getWord16 Motorola = getWord16be getWord16 Intel = getWord16le -- | Little support function to read 32 bit integers getWord32 :: Encoding -> Get Word32 getWord32 Motorola = getWord32be getWord32 Intel = getWord32le -- | Convert a lazy ByteString into a normal Haskell String -- Copied from MissingH library module Data.Bits.Utils unpackLazyBS :: BL.ByteString -> String unpackLazyBS = map (toEnum . fromIntegral) . BL.unpack showHex :: Binary a => a -> String showHex = show . hex . encode runGetEither :: Get a -> BL.ByteString -> Either String a runGetEither get bs = case runGetOrFail get bs of Left (_,_,strError) -> Left strError Right (_,_,x) -> Right x -- Return the first parameter, if it's Nothing, return the second parameter -- This something similar to an OR for Maybe values ifNothing :: Maybe a -> Maybe a -> Maybe a ifNothing mbX mbY = case mbX of Nothing -> mbY _ -> mbX
hansroland/hexif
src/Graphics/Hexif/Utils.hs
bsd-3-clause
1,291
0
9
252
306
165
141
26
2
{-# LANGUAGE Arrows #-} module Karamaan.Plankton.Arrow where import Prelude hiding (id) import Control.Arrow ((&&&), Arrow, arr, returnA, second) import Control.Category ((<<<), id) import Data.List (foldl') voidArr :: Arrow arr => arr a () voidArr = arr (const ()) andVoid :: Arrow arr => arr a () -> arr a () -> arr a () andVoid f g = voidArr <<< (f &&& g) all_ :: Arrow arr => [arr a ()] -> arr a () all_ = foldl' andVoid voidArr replaceWith :: Arrow arr => arr () b -> arr a b replaceWith = (<<< voidArr) restrictWith :: Arrow arr => arr b () -> arr b b restrictWith r = arr fst <<< (id &&& r) removeUnit :: Arrow arr => arr a (a, ()) removeUnit = arr (flip (,) ()) liftArr2 :: Arrow arr => (a -> b -> r) -> arr z a -> arr z b -> arr z r liftArr2 f x y = f <$-> x <*-> y -- TODO: I guess this should really be made to work with any traversal -- cf 'instance Traversable []' in Data.Traversable sequenceArr :: Arrow arr => [arr a b] -> arr a [b] sequenceArr = foldr (liftArr2 (:)) (arr (const [])) traverseArr :: Arrow arr => (b -> arr a c) -> [b] -> arr a [c] traverseArr f = sequenceArr . map f (<*->) :: Arrow arr => arr a (b -> c) -> arr a b -> arr a c f <*-> x = proc a -> do f' <- f -< a x' <- x -< a returnA -< f' x' (<$->) :: Arrow arr => (b -> c) -> arr a b -> arr a c f <$-> x = arr f <<< x foldrArr :: Arrow arr => arr (a, b) b -> arr z b -> [arr z a] -> arr z b foldrArr f = foldr g where g x rest = f <<< (x &&& rest) foldl'Arr :: Arrow arr => arr (b, a) b -> arr z b -> [arr z a] -> arr z b foldl'Arr f = foldl' g where g rest x = f <<< (rest &&& x) -- opC stands for "operator curry". It's a sort of partial application. -- If anyone has a better name, please change it :) opC :: Arrow arr => arr (a, b) c -> arr () b -> arr a c opC op q = op <<< second q <<< arr (\a -> (a, ())) noOp :: Arrow arr => arr a () noOp = arr (const ())
karamaan/karamaan-plankton
Karamaan/Plankton/Arrow.hs
bsd-3-clause
1,878
1
10
459
978
496
482
41
1
module SLM.Model where import SLM.DataTypes class TrainedModel a where predictInstance :: a -> [Predictor] -> Double predict :: a -> [[Predictor]] -> [Double] predict model = map (predictInstance model)
timveitch/Gobble
src/SLM/Model.hs
bsd-3-clause
211
0
10
37
75
41
34
6
0
-- | -- Module : -- License : BSD-Style -- Maintainer : Nicolas DI PRIMA <[email protected]> -- Stability : experimental -- Portability : unknown -- module Network.SMTP ( module Network.SMTP.Types , module Network.SMTP.Monad -- * Client , module Network.SMTP.Client -- * Utils -- ** Lookup MX domains , MXServer(..) , lookupMXs -- ** email address , emailAddrFromString -- ** pretty print result , prettyPrintResult ) where import Data.Byteable import Data.ByteString (ByteString) import qualified Data.ByteString.Char8 as BC import qualified Data.ByteString.Parse as BP (parse, Result(..)) import Data.List (intercalate, sort) import Data.Hourglass import qualified Network.DNS as DNS import Network.SMTP.Types import Network.SMTP.Monad import Network.SMTP.Client import Network.SMTP.Parser data MXServer = MXServer { mxWeight :: Int , mxAddress :: Domain } deriving (Eq, Show) instance Ord MXServer where compare mx1 mx2 = compare (mxWeight mx1) (mxWeight mx2) showMXServer :: MXServer -> String showMXServer (MXServer w d) = "(" ++ show w ++ ")" ++ BC.unpack (toBytes d) -- | This function retrieve the MX Server which are reponsible for the given -- email address lookupMXs :: Domain -> SMTP [MXServer] lookupMXs domain = do seed <- smtpLiftIO "get the DNS Resolver information" $ DNS.makeResolvSeed resolvconf eres <- smtpLiftIO "send the DNS MXs Query" $ DNS.withResolver seed $ \r -> DNS.lookupMX r (toBytes domain) case eres of Left err -> fail $ "unable to collect the MX Domains for " ++ BC.unpack (toBytes domain) ++ " error is: " ++ show err Right v -> do let mxs = sort $ map (\(d, p) -> MXServer p (Domain d)) v smtpReportInfo $ ("collecting MX for " ++ BC.unpack (toBytes domain)) ++ ": " ++ (intercalate " " $ map showMXServer mxs) return mxs where resolvconf :: DNS.ResolvConf resolvconf = DNS.defaultResolvConf emailAddrFromString :: String -> Either String EmailAddress emailAddrFromString str = case BP.parse parseEmailAddress (BC.snoc bs '\0') of BP.ParseFail err -> Left $ "failed to parse: " ++ show bs ++ " " ++ err BP.ParseMore _ -> Left $ "failed to parse: " ++ show bs ++ " not enough bytes" BP.ParseOK _ v -> Right v where bs :: ByteString bs = BC.pack str prettyPrintResult :: String -> Result () -> String prettyPrintResult indentation rs = let (report, header) = case rs of SMTPOK r _ -> (r, "Success") SMTPKO r -> (r, "Failure") in indentation ++ header ++ "\n" ++ (prettyPrintReports (indentation ++ " ") report) prettyPrintReports :: String -> SMTPReports -> String prettyPrintReports indentation rs = intercalate "\n" $ map (prettyPrintReport indentation) (reverse $ smtpReports rs) prettyPrintReport :: String -> SMTPReport -> String prettyPrintReport indentation r = case r of SMTPReportInfo {} -> prettyPrintString indentation time "INF" (smtpReportReport r) SMTPReportError {} -> prettyPrintString indentation time "ERR" (smtpReportReport r) SMTPReportCommand {} -> prettyPrintCommand indentation time (smtpReportCmd r) SMTPReportResponse {} -> prettyPrintResponse indentation time (smtpReportRespCode r) where time :: String time = timePrint "EPOCH.p6" $ smtpReportDate r prettyPrintString :: String -> String -> String -> String -> String prettyPrintString indentation time ty msg = indentation ++ "[" ++ time ++ "][" ++ ty ++ "] " ++ msg prettyPrintCommand :: String -> String -> Command -> String prettyPrintCommand indentation time cmd = indentation ++ "[" ++ time ++ "][CMD] " ++ (BC.unpack $ showCommand cmd) prettyPrintResponse :: String -> String -> Response -> String prettyPrintResponse indentation time resp = intercalate "\n" $ map (((++) prefix) . BC.unpack) $ message resp where prefix :: String prefix = indentation ++ "[" ++ time ++ "][RSP][" ++ show (responseCodeToInt $ code resp) ++ "] "
NicolasDP/hs-smtp
Network/SMTP.hs
bsd-3-clause
4,230
0
21
1,057
1,184
619
565
83
4
{-# LANGUAGE RecordWildCards #-} module Plugin.Jira ( respond , hear , help ) where import Control.Applicative import Control.Lens import Data.Aeson import Data.List (nub) import Data.Maybe import Data.Text (Text) import qualified Data.Text as T -- import Data.Text.Lazy (toStrict) -- import Data.Text.Lazy.Encoding import Network.HTTP.Types import Network.Wreq import Text.Megaparsec hiding (label) import Text.Megaparsec.Text import Types import Util import Web.Slack data Issue = Issue { issueKey :: Text , summary :: Text , description :: Maybe Text , priorityName :: Maybe Text , priorityIcon :: Maybe Text , issueStatus :: Maybe Text , assigneeName :: Maybe Text , creatorName :: Maybe Text , issueUrl :: Maybe Text , commentCount :: Maybe Integer } deriving Show data IssueResponseType = Short | Full instance FromJSON Issue where parseJSON (Object v) = do -- parents let fields = v .: "fields" let priority = fields >>= (.: "priority") -- values issueKey <- v .: "key" summary <- fields >>= (.: "summary") description <- optional $ fields >>= (.: "description") priorityName <- optional $ priority >>= (.: "name") priorityIcon <- optional $ priority >>= (.: "iconUrl") issueStatus <- optional $ fields >>= (.: "status") >>= (.: "name") assigneeName <- optional $ fields >>= (.: "assignee") >>= (.: "displayName") creatorName <- optional $ fields >>= (.: "creator") >>= (.: "displayName") commentCount <- optional $ fields >>= (.: "comment") >>= (.: "total") let issueUrl = Nothing pure Issue{..} parseJSON _ = mempty respond :: BotResponse respond (evt, resp) = maybe (pure ()) parseCmd (command evt) where parseCmd cmd = case trigger cmd of "jira" -> fetchIssues (arguments cmd) resp "help" -> help (arguments cmd) resp _ -> pure () hear :: BotResponse hear (evt, resp) = do fetchShortIssues (T.words $ fullMessage evt) resp help :: [Text] -> OutputResponse -> IO () help args resp = case args of [] -> say ["jira"] -> say _ -> pure () where say = slackWriter resp $ QuotedSimpleMessage "!jira XYZ-123 <XYZ-234 XYZ-345…> :: Displays Jira entries" isIssueKey :: Text -> Bool isIssueKey = isJust . parseMaybe issueKeyParser where issueKeyParser :: Parser String issueKeyParser = some letterChar >> char '-' >> some numberChar getIssues :: [Text] -> [Text] getIssues = nub . map T.toUpper . filter isIssueKey fetchShortIssues :: [Text] -> OutputResponse -> IO () fetchShortIssues message resp = mapM_ (fetchIssue Short resp) $ getIssues message fetchIssues :: [Text] -> OutputResponse -> IO () fetchIssues args resp = mapM_ (fetchIssue Full resp) $ getIssues args fetchIssue :: IssueResponseType -> OutputResponse -> Text -> IO () fetchIssue respType resp issueName = do domain <- T.pack <$> envDefault "jira.atlassian.net" "JIRA_DOMAIN" let uri = T.concat ["https://", domain, "/rest/api/2/issue/", issueName] username <- fmap T.pack <$> envMaybe "JIRA_USERNAME" password <- fmap T.pack <$> envMaybe "JIRA_PASSWORD" response <- safeGetUrl uri username password -- TODO thread those guys case response of Left err -> print $ T.concat ["[JIRA] Error: ", err, " :: issue ", issueName] Right success -> do print $ T.concat ["[JIRA] Fetched issue ", issueName] -- print $ toStrict $ decodeUtf8 $ success ^. responseBody let status = success ^. responseStatus let body = decode (success ^. responseBody) :: Maybe Issue case body of Just issue -> slackWriter resp $ handle respType issueName (withUrl domain issue) status Nothing -> pure () withUrl :: Text -> Issue -> Issue withUrl domain issue = issue { issueUrl = Just $ T.concat ["https://", domain, "/browse/", issueKey issue]} handle :: IssueResponseType -> Text -> Issue -> Status -> OutputMessage handle respType issueName issue status | statusIsServerError status = showError issueName | statusIsClientError status = showNotFound issueName | otherwise = case respType of Full -> showIssue issue Short -> showShortIssue issue makeField :: Text -> Maybe Text -> Maybe Field makeField label val = case val of Just v -> Just $ Field (Just label) v True Nothing -> Nothing showShortIssue :: Issue -> OutputMessage showShortIssue issue = QuotedSimpleMessage $ T.intercalate " :: " $ nub $ sequence [issueKey, summary, (fromMaybe "") . issueStatus] issue showIssue :: Issue -> OutputMessage showIssue issue = RichMessage defaultAttachment { attachmentFallback = summary issue , attachmentColor = color , attachmentAuthorName = Just (issueKey issue) , attachmentAuthorIcon = priorityIcon issue , attachmentTitle = Just (summary issue) , attachmentTitleLink = issueUrl issue , attachmentText = description issue , attachmentFields = catMaybes [ makeField "Status" (issueStatus issue) , makeField "Assigned to" (assigneeName issue) ] , attachmentThumbUrl = Just "https://a.slack-edge.com/2fac/plugins/jira/assets/service_36.png" , attachmentFooter = Just $ T.concat ["Comments: ", T.pack . show $ (fromMaybe 0 $ commentCount issue)] } where color = case (priorityName issue) of Just "Blocker" -> DangerColor Just "Critical" -> DangerColor Just "Major" -> WarningColor Just "Minor" -> GoodColor Just "Trivial" -> GoodColor _ -> DefaultColor showError :: Text -> OutputMessage showError issueName = SimpleMessage $ T.concat ["Error fetching item `", issueName, "` from Jira."] showNotFound :: Text -> OutputMessage showNotFound issueName = SimpleMessage $ T.concat ["Item `", issueName, "` not found in Jira."]
wamaral/slaskellbot
src/Plugin/Jira.hs
bsd-3-clause
6,218
0
18
1,663
1,772
914
858
127
6
module Data.Wright.RGB.Model.AdobeRGB1998 (adobeRGB1998) where import Data.Wright.Types (Model(..), Primary(..), Gamma(..)) import Data.Wright.CIE.Illuminant.D65 (d65) adobeRGB1998 :: Model adobeRGB1998 = d65 { gamma = Gamma 2.2 , red = Primary 0.6400 0.3300 0.297361 , green = Primary 0.2100 0.7100 0.627355 , blue = Primary 0.1500 0.0600 0.075285 }
fmap-archive/wright
src/Data/Wright/RGB/Model/AdobeRGB1998.hs
mit
365
0
7
59
116
72
44
9
1
import Data.Array import Data.Function (on) import Data.List (intercalate, maximumBy) import Data.Ratio type Matrix a = Array (Int, Int) a type Vector a = Array Int a swapRows :: Int -> Int -> Matrix a -> Matrix a swapRows r1 r2 m | r1 == r2 = m | otherwise = m // concat [[((r2, c), m ! (r1, c)), ((r1, c), m ! (r2, c))] | c <- [c1 .. cn]] where ((_, c1), (_, cn)) = bounds m subRows :: Fractional a => (Int, Int) -- pivot location -> (Int, Int) -- rows to cover -> (Int, Int) -- columns to cover -> Matrix a -> Matrix a subRows (r, c) (r1, rn) (c1, cn) m = accum (-) m [ ((i, j), m ! (i, c) * m ! (r, j) / m ! (r, c)) | i <- [r1 .. rn] , j <- [c1 .. cn] ] gaussianElimination :: (Fractional a, Ord a) => Matrix a -> Matrix a gaussianElimination mat = go (r1, c1) mat where ((r1, c1), (rn, cn)) = bounds mat go (r, c) m | c == cn = m | pivot == 0 = go (r, c + 1) m | otherwise = go (r + 1, c + 1) $ subRows (r, c) (r + 1, rn) (c, cn) m' where (target, pivot) = maximumBy (compare `on` abs . snd) [(k, m ! (k, c)) | k <- [r .. rn]] m' = swapRows r target m gaussJordan :: (Fractional a, Eq a) => Matrix a -> Matrix a gaussJordan mat = go (r1, c1) mat where ((r1, c1), (rn, cn)) = bounds mat go (r, c) m | c == cn = m | m ! (r, c) == 0 = go (r, c + 1) m | otherwise = go (r + 1, c + 1) $ subRows (r, c) (r1, r - 1) (c, cn) m' where m' = accum (/) m [((r, j), m ! (r, c)) | j <- [c .. cn]] backSubstitution :: (Fractional a) => Matrix a -> Vector a backSubstitution m = sol where ((r1, _), (rn, cn)) = bounds m sol = listArray (r1, rn) [(m ! (r, cn) - sum' r) / m ! (r, r) | r <- [r1 .. rn]] sum' r = sum [m ! (r, k) * sol ! k | k <- [r + 1 .. rn]] printM :: (Show a) => Matrix a -> String printM m = let ((r1, c1), (rn, cn)) = bounds m in unlines [ intercalate "\t" [show $ m ! (r, c) | c <- [c1 .. cn]] | r <- [r1 .. rn] ] printV :: (Show a) => Vector a -> String printV = unlines . map show . elems main :: IO () main = do let mat = [2, 3, 4, 6, 1, 2, 3, 4, 3, -4, 0, 10] :: [Ratio Int] m = listArray ((1, 1), (3, 4)) mat putStrLn "Original Matrix:" putStrLn $ printM m putStrLn "Echelon form" putStrLn $ printM $ gaussianElimination m putStrLn "Reduced echelon form" putStrLn $ printM $ gaussJordan $ gaussianElimination m putStrLn "Solution from back substitution" putStrLn $ printV $ backSubstitution $ gaussianElimination m
Gathros/algorithm-archive
contents/gaussian_elimination/code/haskell/gaussianElimination.hs
mit
2,573
10
12
781
1,495
798
697
-1
-1
{-# LANGUAGE QuasiQuotes #-} {-# LANGUAGE TemplateHaskell #-} {-# LANGUAGE TypeFamilies #-} {-# LANGUAGE CPP #-} {-# LANGUAGE FlexibleInstances #-} {-# LANGUAGE MultiParamTypeClasses #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE RecordWildCards #-} {-# LANGUAGE FlexibleContexts #-} --------------------------------------------------------- -- -- | Serve static files from a Yesod app. -- -- This is great for developing your application, but also for a -- dead-simple deployment. Caching headers are automatically -- taken care of. -- -- If you are running a proxy server (like Apache or Nginx), -- you may want to have that server do the static serving instead. -- -- In fact, in an ideal setup you'll serve your static files from -- a separate domain name to save time on transmitting -- cookies. In that case, you may wish to use 'urlParamRenderOverride' -- to redirect requests to this subsite to a separate domain -- name. -- -- Note that this module's static subsite ignores all files and -- directories that are hidden by Unix conventions (i.e. start -- with a dot, such as @\".ssh\"@) and the directory "tmp" on the -- root of the directory with static files. module Yesod.Static ( -- * Subsite Static (..) , Route (..) , StaticRoute -- * Smart constructor , static , staticDevel -- * Combining CSS/JS -- $combining , combineStylesheets' , combineScripts' -- ** Settings , CombineSettings , csStaticDir , csCssPostProcess , csJsPostProcess , csCssPreProcess , csJsPreProcess , csCombinedFolder -- * Template Haskell helpers , staticFiles , staticFilesList , staticFilesMap , staticFilesMergeMap , publicFiles -- * Hashing , base64md5 -- * Embed , embed #ifdef TEST_EXPORT , getFileListPieces #endif ) where import System.Directory import qualified System.FilePath as FP import Control.Monad import Data.FileEmbed (embedDir) import Yesod.Core import Yesod.Core.Types import Data.List (intercalate, sort) import Language.Haskell.TH import Language.Haskell.TH.Syntax as TH import Crypto.Hash.Conduit (hashFile, sinkHash) import Crypto.Hash (MD5, Digest) import Control.Monad.Trans.State import qualified Data.ByteArray as ByteArray import qualified Data.ByteString.Base64 import qualified Data.ByteString.Char8 as S8 import qualified Data.ByteString.Lazy as L import Data.Text (Text, pack) import qualified Data.Text as T import qualified Data.Map as M import Data.IORef (readIORef, newIORef, writeIORef) import Data.Char (isLower, isDigit) import Data.List (foldl') import qualified Data.ByteString as S import System.PosixCompat.Files (getFileStatus, modificationTime) import System.Posix.Types (EpochTime) import Conduit import System.FilePath ((</>), (<.>), takeDirectory) import qualified System.FilePath as F import qualified Data.Text.Lazy as TL import qualified Data.Text.Lazy.Encoding as TLE import Data.Default --import Text.Lucius (luciusRTMinified) import Network.Wai (pathInfo) import Network.Wai.Application.Static ( StaticSettings (..) , staticApp , webAppSettingsWithLookup , embeddedSettings ) import WaiAppStatic.Storage.Filesystem (ETagLookup) -- | Type used for the subsite with static contents. newtype Static = Static StaticSettings type StaticRoute = Route Static -- | Produce a default value of 'Static' for a given file -- folder. -- -- Does not have index files or directory listings. The static -- files' contents /must not/ change, however new files can be -- added. static :: FilePath -> IO Static static dir = do hashLookup <- cachedETagLookup dir return $ Static $ webAppSettingsWithLookup dir hashLookup -- | Same as 'static', but does not assumes that the files do not -- change and checks their modification time whenever a request -- is made. staticDevel :: FilePath -> IO Static staticDevel dir = do hashLookup <- cachedETagLookupDevel dir return $ Static $ webAppSettingsWithLookup dir hashLookup -- | Produce a 'Static' based on embedding all of the static files' contents in the -- executable at compile time. -- -- You should use "Yesod.EmbeddedStatic" instead, it is much more powerful. -- -- Nota Bene: if you replace the scaffolded 'static' call in Settings/StaticFiles.hs -- you will need to change the scaffolded addStaticContent. Otherwise, some of your -- assets will be 404'ed. This is because by default yesod will generate compile those -- assets to @static/tmp@ which for 'static' is fine since they are served out of the -- directory itself. With embedded static, that will not work. -- You can easily change @addStaticContent@ to @\_ _ _ -> return Nothing@ as a workaround. -- This will cause yesod to embed those assets into the generated HTML file itself. embed :: FilePath -> Q Exp embed fp = [|Static (embeddedSettings $(embedDir fp))|] instance RenderRoute Static where -- | A route on the static subsite (see also 'staticFiles'). -- -- You may use this constructor directly to manually link to a -- static file. The first argument is the sub-path to the file -- being served whereas the second argument is the key-value -- pairs in the query string. For example, -- -- > StaticRoute $ StaticR [\"thumb001.jpg\"] [(\"foo\", \"5\"), (\"bar\", \"choc\")] -- -- would generate a url such as -- @http://www.example.com/static/thumb001.jpg?foo=5&bar=choc@ -- The StaticRoute constructor can be used when the URL cannot be -- statically generated at compile-time (e.g. when generating -- image galleries). data Route Static = StaticRoute [Text] [(Text, Text)] deriving (Eq, Show, Read) renderRoute (StaticRoute x y) = (x, y) instance ParseRoute Static where parseRoute (x, y) = Just $ StaticRoute x y instance YesodSubDispatch Static master where yesodSubDispatch YesodSubRunnerEnv {..} req = ysreParentRunner handlert ysreParentEnv (fmap ysreToParentRoute route) req where route = Just $ StaticRoute (pathInfo req) [] Static set = ysreGetSub $ yreSite $ ysreParentEnv handlert = sendWaiApplication $ staticApp set notHidden :: FilePath -> Bool notHidden "tmp" = False notHidden s = case s of '.':_ -> False _ -> True getFileListPieces :: FilePath -> IO [[String]] getFileListPieces = flip evalStateT M.empty . flip go id where go :: String -> ([String] -> [String]) -> StateT (M.Map String String) IO [[String]] go fp front = do allContents <- liftIO $ (sort . filter notHidden) `fmap` getDirectoryContents fp let fullPath :: String -> String fullPath f = fp ++ '/' : f files <- liftIO $ filterM (doesFileExist . fullPath) allContents let files' = map (front . return) files files'' <- mapM dedupe files' dirs <- liftIO $ filterM (doesDirectoryExist . fullPath) allContents dirs' <- mapM (\f -> go (fullPath f) (front . (:) f)) dirs return $ concat $ files'' : dirs' -- Reuse data buffers for identical strings dedupe :: [String] -> StateT (M.Map String String) IO [String] dedupe = mapM dedupe' dedupe' :: String -> StateT (M.Map String String) IO String dedupe' s = do m <- get case M.lookup s m of Just s' -> return s' Nothing -> do put $ M.insert s s m return s -- | Template Haskell function that automatically creates routes -- for all of your static files. -- -- For example, if you used -- -- > staticFiles "static/" -- -- and you had files @\"static\/style.css\"@ and -- @\"static\/js\/script.js\"@, then the following top-level -- definitions would be created: -- -- > style_css = StaticRoute ["style.css"] [] -- > js_script_js = StaticRoute ["js", "script.js"] [] -- -- Note that dots (@.@), dashes (@-@) and slashes (@\/@) are -- replaced by underscores (@\_@) to create valid Haskell -- identifiers. staticFiles :: FilePath -> Q [Dec] staticFiles dir = mkStaticFiles dir -- | Same as 'staticFiles', but takes an explicit list of files -- to create identifiers for. The files path given are relative -- to the static folder. For example, to create routes for the -- files @\"static\/js\/jquery.js\"@ and -- @\"static\/css\/normalize.css\"@, you would use: -- -- > staticFilesList "static" ["js/jquery.js", "css/normalize.css"] -- -- This can be useful when you have a very large number of static -- files, but only need to refer to a few of them from Haskell. staticFilesList :: FilePath -> [FilePath] -> Q [Dec] staticFilesList dir fs = mkStaticFilesList dir (map split fs) True where split :: FilePath -> [String] split [] = [] split x = let (a, b) = break (== '/') x in a : split (drop 1 b) -- | Same as 'staticFiles', but doesn't append an ETag to the -- query string. -- -- Using 'publicFiles' will speed up the compilation, since there -- won't be any need for hashing files during compile-time. -- However, since the ETag ceases to be part of the URL, the -- 'Static' subsite won't be able to set the expire date too far -- on the future. Browsers still will be able to cache the -- contents, however they'll need send a request to the server to -- see if their copy is up-to-date. publicFiles :: FilePath -> Q [Dec] publicFiles dir = mkStaticFiles' dir False -- | Similar to 'staticFilesList', but takes a mapping of -- unmunged names to fingerprinted file names. -- -- @since 1.5.3 staticFilesMap :: FilePath -> M.Map FilePath FilePath -> Q [Dec] staticFilesMap fp m = mkStaticFilesList' fp (map splitBoth mapList) True where splitBoth (k, v) = (split k, split v) mapList = M.toList m split :: FilePath -> [String] split [] = [] split x = let (a, b) = break (== '/') x in a : split (drop 1 b) -- | Similar to 'staticFilesMergeMap', but also generates identifiers -- for all files in the specified directory that don't have a -- fingerprinted version. -- -- @since 1.5.3 staticFilesMergeMap :: FilePath -> M.Map FilePath FilePath -> Q [Dec] staticFilesMergeMap fp m = do fs <- qRunIO $ getFileListPieces fp let filesList = map FP.joinPath fs mergedMapList = M.toList $ foldl' (checkedInsert invertedMap) m filesList mkStaticFilesList' fp (map splitBoth mergedMapList) True where splitBoth (k, v) = (split k, split v) swap (x, y) = (y, x) mapList = M.toList m invertedMap = M.fromList $ map swap mapList split :: FilePath -> [String] split [] = [] split x = let (a, b) = break (== '/') x in a : split (drop 1 b) -- We want to keep mappings for all files that are pre-fingerprinted, -- so this function checks against all of the existing fingerprinted files and -- only inserts a new mapping if it's not a fingerprinted file. checkedInsert :: M.Map FilePath FilePath -- inverted dictionary -> M.Map FilePath FilePath -- accumulating state -> FilePath -> M.Map FilePath FilePath checkedInsert iDict st p = if M.member p iDict then st else M.insert p p st mkHashMap :: FilePath -> IO (M.Map FilePath S8.ByteString) mkHashMap dir = do fs <- getFileListPieces dir hashAlist fs >>= return . M.fromList where hashAlist :: [[String]] -> IO [(FilePath, S8.ByteString)] hashAlist fs = mapM hashPair fs where hashPair :: [String] -> IO (FilePath, S8.ByteString) hashPair pieces = do let file = pathFromRawPieces dir pieces h <- base64md5File file return (file, S8.pack h) pathFromRawPieces :: FilePath -> [String] -> FilePath pathFromRawPieces = foldl' append where append a b = a ++ '/' : b cachedETagLookupDevel :: FilePath -> IO ETagLookup cachedETagLookupDevel dir = do etags <- mkHashMap dir mtimeVar <- newIORef (M.empty :: M.Map FilePath EpochTime) return $ \f -> case M.lookup f etags of Nothing -> return Nothing Just checksum -> do fs <- getFileStatus f let newt = modificationTime fs mtimes <- readIORef mtimeVar oldt <- case M.lookup f mtimes of Nothing -> writeIORef mtimeVar (M.insert f newt mtimes) >> return newt Just oldt -> return oldt return $ if newt /= oldt then Nothing else Just checksum cachedETagLookup :: FilePath -> IO ETagLookup cachedETagLookup dir = do etags <- mkHashMap dir return $ (\f -> return $ M.lookup f etags) mkStaticFiles :: FilePath -> Q [Dec] mkStaticFiles fp = mkStaticFiles' fp True mkStaticFiles' :: FilePath -- ^ static directory -> Bool -- ^ append checksum query parameter -> Q [Dec] mkStaticFiles' fp makeHash = do fs <- qRunIO $ getFileListPieces fp mkStaticFilesList fp fs makeHash mkStaticFilesList :: FilePath -- ^ static directory -> [[String]] -- ^ list of files to create identifiers for -> Bool -- ^ append checksum query parameter -> Q [Dec] mkStaticFilesList fp fs makeHash = mkStaticFilesList' fp (zip fs fs) makeHash mkStaticFilesList' :: FilePath -- ^ static directory -> [([String], [String])] -- ^ list of files to create identifiers for, where -- the first argument of the tuple is the identifier -- alias and the second is the actual file name -> Bool -- ^ append checksum query parameter -> Q [Dec] mkStaticFilesList' fp fs makeHash = do concat `fmap` mapM mkRoute fs where replace' c | 'A' <= c && c <= 'Z' = c | 'a' <= c && c <= 'z' = c | '0' <= c && c <= '9' = c | otherwise = '_' mkRoute (alias, f) = do let name' = intercalate "_" $ map (map replace') alias routeName = mkName $ case () of () | null name' -> error "null-named file" | isDigit (head name') -> '_' : name' | isLower (head name') -> name' | otherwise -> '_' : name' f' <- [|map pack $(TH.lift f)|] qs <- if makeHash then do hash <- qRunIO $ base64md5File $ pathFromRawPieces fp f [|[(pack "etag", pack $(TH.lift hash))]|] else return $ ListE [] return [ SigD routeName $ ConT ''StaticRoute , FunD routeName [ Clause [] (NormalB $ (ConE 'StaticRoute) `AppE` f' `AppE` qs) [] ] ] base64md5File :: FilePath -> IO String base64md5File = fmap (base64 . encode) . hashFile where encode d = ByteArray.convert (d :: Digest MD5) base64md5 :: L.ByteString -> String base64md5 lbs = base64 $ encode $ runConduitPure $ Conduit.sourceLazy lbs .| sinkHash where encode d = ByteArray.convert (d :: Digest MD5) base64 :: S.ByteString -> String base64 = map tr . take 8 . S8.unpack . Data.ByteString.Base64.encode where tr '+' = '-' tr '/' = '_' tr c = c -- $combining -- -- A common scenario on a site is the desire to include many external CSS and -- Javascript files on every page. Doing so via the Widget functionality in -- Yesod will work, but would also mean that the same content will be -- downloaded many times. A better approach would be to combine all of these -- files together into a single static file and serve that as a static resource -- for every page. That resource can be cached on the client, and bandwidth -- usage reduced. -- -- This could be done as a manual process, but that becomes tedious. Instead, -- you can use some Template Haskell code which will combine these files into a -- single static file at compile time. data CombineType = JS | CSS combineStatics' :: CombineType -> CombineSettings -> [Route Static] -- ^ files to combine -> Q Exp combineStatics' combineType CombineSettings {..} routes = do texts <- qRunIO $ runConduitRes $ yieldMany fps .| awaitForever readUTFFile .| sinkLazy ltext <- qRunIO $ preProcess texts bs <- qRunIO $ postProcess fps $ TLE.encodeUtf8 ltext let hash' = base64md5 bs suffix = csCombinedFolder </> hash' <.> extension fp = csStaticDir </> suffix qRunIO $ do createDirectoryIfMissing True $ takeDirectory fp L.writeFile fp bs let pieces = map T.unpack $ T.splitOn "/" $ T.pack suffix [|StaticRoute (map pack pieces) []|] where fps :: [FilePath] fps = map toFP routes toFP (StaticRoute pieces _) = csStaticDir </> F.joinPath (map T.unpack pieces) readUTFFile fp = sourceFile fp .| decodeUtf8C postProcess = case combineType of JS -> csJsPostProcess CSS -> csCssPostProcess preProcess = case combineType of JS -> csJsPreProcess CSS -> csCssPreProcess extension = case combineType of JS -> "js" CSS -> "css" -- | Data type for holding all settings for combining files. -- -- This data type is a settings type. For more information, see: -- -- <http://www.yesodweb.com/book/settings-types> -- -- Since 1.2.0 data CombineSettings = CombineSettings { csStaticDir :: FilePath -- ^ File path containing static files. -- -- Default: static -- -- Since 1.2.0 , csCssPostProcess :: [FilePath] -> L.ByteString -> IO L.ByteString -- ^ Post processing to be performed on CSS files. -- -- Default: Pass-through. -- -- Since 1.2.0 , csJsPostProcess :: [FilePath] -> L.ByteString -> IO L.ByteString -- ^ Post processing to be performed on Javascript files. -- -- Default: Pass-through. -- -- Since 1.2.0 , csCssPreProcess :: TL.Text -> IO TL.Text -- ^ Pre-processing to be performed on CSS files. -- -- Default: convert all occurences of /static/ to ../ -- -- Since 1.2.0 , csJsPreProcess :: TL.Text -> IO TL.Text -- ^ Pre-processing to be performed on Javascript files. -- -- Default: Pass-through. -- -- Since 1.2.0 , csCombinedFolder :: FilePath -- ^ Subfolder to put combined files into. -- -- Default: combined -- -- Since 1.2.0 } instance Default CombineSettings where def = CombineSettings { csStaticDir = "static" {- Disabled due to: https://github.com/yesodweb/yesod/issues/623 , csCssPostProcess = \fps -> either (error . (errorIntro fps)) (return . TLE.encodeUtf8) . flip luciusRTMinified [] . TLE.decodeUtf8 -} , csCssPostProcess = const return , csJsPostProcess = const return -- FIXME The following borders on a hack. With combining of files, -- the final location of the CSS is no longer fixed, so relative -- references will break. Instead, we switched to using /static/ -- absolute references. However, when served from a separate domain -- name, this will break too. The solution is that, during -- development, we keep /static/, and in the combining phase, we -- replace /static with a relative reference to the parent folder. , csCssPreProcess = return . TL.replace "'/static/" "'../" . TL.replace "\"/static/" "\"../" , csJsPreProcess = return , csCombinedFolder = "combined" } liftRoutes :: [Route Static] -> Q Exp liftRoutes = fmap ListE . mapM go where go :: Route Static -> Q Exp go (StaticRoute x y) = [|StaticRoute $(liftTexts x) $(liftPairs y)|] liftTexts = fmap ListE . mapM liftT liftT t = [|pack $(TH.lift $ T.unpack t)|] liftPairs = fmap ListE . mapM liftPair liftPair (x, y) = [|($(liftT x), $(liftT y))|] -- | Combine multiple CSS files together. Common usage would be: -- -- >>> combineStylesheets' development def 'StaticR [style1_css, style2_css] -- -- Where @development@ is a variable in your site indicated whether you are in -- development or production mode. -- -- Since 1.2.0 combineStylesheets' :: Bool -- ^ development? if so, perform no combining -> CombineSettings -> Name -- ^ Static route constructor name, e.g. \'StaticR -> [Route Static] -- ^ files to combine -> Q Exp combineStylesheets' development cs con routes | development = [| mapM_ (addStylesheet . $(return $ ConE con)) $(liftRoutes routes) |] | otherwise = [| addStylesheet $ $(return $ ConE con) $(combineStatics' CSS cs routes) |] -- | Combine multiple JS files together. Common usage would be: -- -- >>> combineScripts' development def 'StaticR [script1_js, script2_js] -- -- Where @development@ is a variable in your site indicated whether you are in -- development or production mode. -- -- Since 1.2.0 combineScripts' :: Bool -- ^ development? if so, perform no combining -> CombineSettings -> Name -- ^ Static route constructor name, e.g. \'StaticR -> [Route Static] -- ^ files to combine -> Q Exp combineScripts' development cs con routes | development = [| mapM_ (addScript . $(return $ ConE con)) $(liftRoutes routes) |] | otherwise = [| addScript $ $(return $ ConE con) $(combineStatics' JS cs routes) |]
geraldus/yesod
yesod-static/Yesod/Static.hs
mit
21,599
0
22
5,540
4,118
2,244
1,874
342
4
{-# LANGUAGE DeriveDataTypeable #-} {-# LANGUAGE DeriveGeneric #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE RecordWildCards #-} {-# LANGUAGE TypeFamilies #-} {-# OPTIONS_GHC -fno-warn-unused-imports #-} {-# OPTIONS_GHC -fno-warn-unused-binds #-} {-# OPTIONS_GHC -fno-warn-unused-matches #-} -- Derived from AWS service descriptions, licensed under Apache 2.0. -- | -- Module : Network.AWS.Lambda.DeleteEventSourceMapping -- Copyright : (c) 2013-2015 Brendan Hay -- License : Mozilla Public License, v. 2.0. -- Maintainer : Brendan Hay <[email protected]> -- Stability : auto-generated -- Portability : non-portable (GHC extensions) -- -- Removes an event source mapping. This means AWS Lambda will no longer -- invoke the function for events in the associated source. -- -- This operation requires permission for the -- 'lambda:DeleteEventSourceMapping' action. -- -- /See:/ <http://docs.aws.amazon.com/lambda/latest/dg/API_DeleteEventSourceMapping.html AWS API Reference> for DeleteEventSourceMapping. module Network.AWS.Lambda.DeleteEventSourceMapping ( -- * Creating a Request deleteEventSourceMapping , DeleteEventSourceMapping -- * Request Lenses , desmUUId -- * Destructuring the Response , eventSourceMappingConfiguration , EventSourceMappingConfiguration -- * Response Lenses , esmcEventSourceARN , esmcState , esmcFunctionARN , esmcUUId , esmcLastProcessingResult , esmcBatchSize , esmcStateTransitionReason , esmcLastModified ) where import Network.AWS.Lambda.Types import Network.AWS.Lambda.Types.Product import Network.AWS.Prelude import Network.AWS.Request import Network.AWS.Response -- | /See:/ 'deleteEventSourceMapping' smart constructor. newtype DeleteEventSourceMapping = DeleteEventSourceMapping' { _desmUUId :: Text } deriving (Eq,Read,Show,Data,Typeable,Generic) -- | Creates a value of 'DeleteEventSourceMapping' with the minimum fields required to make a request. -- -- Use one of the following lenses to modify other fields as desired: -- -- * 'desmUUId' deleteEventSourceMapping :: Text -- ^ 'desmUUId' -> DeleteEventSourceMapping deleteEventSourceMapping pUUId_ = DeleteEventSourceMapping' { _desmUUId = pUUId_ } -- | The event source mapping ID. desmUUId :: Lens' DeleteEventSourceMapping Text desmUUId = lens _desmUUId (\ s a -> s{_desmUUId = a}); instance AWSRequest DeleteEventSourceMapping where type Rs DeleteEventSourceMapping = EventSourceMappingConfiguration request = delete lambda response = receiveJSON (\ s h x -> eitherParseJSON x) instance ToHeaders DeleteEventSourceMapping where toHeaders = const mempty instance ToPath DeleteEventSourceMapping where toPath DeleteEventSourceMapping'{..} = mconcat ["/2015-03-31/event-source-mappings/", toBS _desmUUId] instance ToQuery DeleteEventSourceMapping where toQuery = const mempty
fmapfmapfmap/amazonka
amazonka-lambda/gen/Network/AWS/Lambda/DeleteEventSourceMapping.hs
mpl-2.0
3,080
0
9
621
336
209
127
53
1
{-# LANGUAGE TemplateHaskell #-} {-# LANGUAGE OverloadedStrings #-} module BitcoinCore.Transaction.Transactions where import General.Util import BitcoinCore.Transaction.Script import BitcoinCore.Keys (PublicKeyRep(..), PubKeyFormat(..)) import General.Hash (Hash(..), hashObject, doubleSHA) import Prelude hiding (concat, reverse, sequence) import Data.ByteString (ByteString) import qualified Data.ByteString as BS import qualified Data.ByteString.Lazy as BL import Crypto.PubKey.ECC.ECDSA ( signWith , Signature(..) , PrivateKey(..) , PublicKey(..) ) import Crypto.Hash.Algorithms (SHA256(..)) import Control.Lens (makeLenses, (^.), mapped, set) import qualified Data.Binary as BIN import qualified Data.Binary.Put as Put import Data.Binary.Put (Put) import qualified Data.Binary.Get as Get import Data.Binary.Get (Get) import Data.Binary (Binary(..), Word32) import Control.Monad (replicateM) import Data.Bits ((.&.)) import Crypto.Hash (hashWith) import Data.ByteArray (convert) import Data.Maybe (fromMaybe) import Test.QuickCheck.Arbitrary (Arbitrary(..)) import Test.QuickCheck.Gen (choose, suchThat) data Transaction = Transaction { _inputs :: [TxInput] , _outputs :: [TxOutput] , _txVersion :: TxVersion , _locktime :: LockTime } deriving (Eq, Show) data TxInput = TxInput { _utxo :: UTXO , _signatureScript :: Script , _sequence :: Sequence } deriving (Eq, Show) data TxOutput = TxOutput { _value :: Value , _outputScript :: Script } deriving (Eq, Show) data UTXO = UTXO { _outTxHash :: TxHash , _outIndex :: TxIndex } deriving (Eq, Show) newtype Value = Satoshis Int deriving (Eq, Show) newtype TxVersion = TxVersion Int deriving (Eq, Show) type TxHash = Hash Transaction newtype TxIndex = TxIndex Int deriving (Eq, Show) newtype Sequence = Sequence Word32 deriving (Eq, Show) newtype LockTime = LockTime Word32 deriving (Eq, Show) makeLenses ''Transaction makeLenses ''TxInput makeLenses ''TxOutput makeLenses ''UTXO hashTransaction :: Transaction -> Hash Transaction hashTransaction = hashObject doubleSHA outputScripts :: Transaction -> [Script] outputScripts transaction = map (^.outputScript) (transaction^.outputs) -------------------- Transaction signing signedTransaction :: UTXO -> Script -> (PublicKey, PrivateKey) -> [TxOutput] -> Transaction signedTransaction utxo' oldInputScript keys outputs' = set (inputs.mapped.signatureScript) newInputScript transaction where newInputScript = scriptSig (signedHash oldInputScript (snd keys) transaction) (fst keys) transaction = Transaction {_inputs = [TxInput { _utxo = utxo' , _sequence = defaultSequence}] , _outputs = outputs' , _txVersion = TxVersion 1 , _locktime = defaultLockTime} -- We need to sign the double SHA the intermediateTransaction. -- Since `signWith` always performs a SHA, we achieve correct -- behaviour by performing one SHA in `intermediateHash` -- and allowing `signWith` to perform the second hash signedHash :: Script -> PrivateKey -> Transaction -> Signature signedHash oldInputScript privateKey intermediateTransaction = fromMaybe (error "Unable to sign hash") (signWith 100 privateKey SHA256 intermediateHash') where intermediateHash' = intermediateHash intermediateTransaction oldInputScript intermediateHash :: Transaction -> Script -> ByteString intermediateHash intermediateTransaction oldInputScript = convert . hashWith SHA256 $ bs where bs = BL.toStrict . Put.runPut $ do put (set (inputs.mapped.signatureScript) oldInputScript intermediateTransaction) Put.putWord32le sighashAll sighashAll :: Word32 sighashAll = 0x00000001 instance Binary Transaction where put = putTransaction get = getTransaction putTransaction :: Transaction -> Put putTransaction tx = do put (tx^.txVersion) put . VarInt . fromIntegral . length $ (tx^.inputs) mapM_ put (tx^.inputs) put . VarInt . fromIntegral . length $ (tx^.outputs) mapM_ put (tx^.outputs) put $ tx^.locktime getTransaction :: Get Transaction getTransaction = do v <- get VarInt inputCount <- get inputArray <- replicateM (fromIntegral inputCount) get VarInt outputCount <- get outputArray <- replicateM (fromIntegral outputCount) get locktime' <- get return Transaction { _inputs = inputArray , _outputs = outputArray , _txVersion = v , _locktime = locktime'} instance Binary TxInput where put = putInput get = getInput putInput :: TxInput -> Put putInput txInput = do put (txInput^.utxo) putWithLength (putScript (txInput^.signatureScript)) put $ txInput^.sequence getInput :: Get TxInput getInput = do outPoint <- get VarInt scriptLength <- get script <- getScript (fromIntegral scriptLength) sequence' <- get return TxInput { _utxo = outPoint , _signatureScript = script , _sequence = sequence' } instance Binary TxOutput where put = putOutput get = getOutput putOutput :: TxOutput -> Put putOutput txOutput = do put (txOutput^.value) putWithLength (putScript (txOutput^.outputScript)) getOutput :: Get TxOutput getOutput = do val <- get VarInt scriptLength <- get script <- getScript (fromIntegral scriptLength) return TxOutput { _value = val , _outputScript = script } instance Binary UTXO where put = putOutPoint get = getOutPoint putOutPoint :: UTXO -> Put putOutPoint utxo' = do put (utxo'^.outTxHash) let TxIndex i = utxo'^.outIndex Put.putWord32le . fromIntegral $ i getOutPoint :: Get UTXO getOutPoint = UTXO <$> get <*> (TxIndex . fromIntegral <$> Get.getWord32le) instance Binary Value where put = putTxValue get = getTxValue putTxValue :: Value -> Put putTxValue (Satoshis i) = Put.putWord64le . fromIntegral $ i getTxValue :: Get Value getTxValue = Satoshis . fromIntegral <$> Get.getWord64le scriptSig :: Signature -> PublicKey -> Script scriptSig signature pubKey = Script [Txt der, Txt compressedPubkey] where der = BL.toStrict . Put.runPut $ putDerSignature signature compressedPubkey = BL.toStrict . BIN.encode $ PublicKeyRep Compressed pubKey -- TODO: This scriptSig will only be valid for -- pay to pub key hash scripts where the compressed pub key is hashed -- make sure that I'm making addresses using compressed pubkeys also -- See https://github.com/bitcoin/bips/blob/master/bip-0066.mediawiki -- for a description of requiered der format putDerSignature :: Signature -> Put putDerSignature signature = do Put.putWord8 0x30 putWithLength (putDERContents signature) Put.putWord8 0x01 -- one byte hashcode type putDERInt :: Integer -> Put putDERInt int = do let intBS = unroll BE int headByte = BS.head intBS if headByte .&. 0x80 == 0x80 then Put.putByteString $ 0x00 `BS.cons` intBS else Put.putByteString intBS putDERContents signature = do Put.putWord8 0x02 putWithLength (putDERInt . sign_r $ signature) Put.putWord8 0x02 putWithLength (putDERInt . getLowS . sign_s $ signature) -- Multiple s values can yield the same signature -- to prevent transaction malleability, s values are required to use the "low s" value -- See: https://github.com/bitcoin/bips/blob/master/bip-0062.mediawiki#low-s-values-in-signatures getLowS :: Integer -> Integer getLowS s = if s <= maxS then s else constant - s where maxS = 0x7FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF5D576E7357A4501DDFE92F46681B20A0 constant = 0xFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFEBAAEDCE6AF48A03BBFD25E8CD0364141 getDerSignature :: Get Signature getDerSignature = do sequenceCode <- Get.getWord8 derLength <- fromIntegral <$> Get.getWord8 Get.getWord8 xLength <- fromIntegral <$> Get.getWord8 x <- roll BE <$> Get.getByteString xLength Get.getWord8 yLength <- fromIntegral <$> Get.getWord8 y <- roll BE <$> Get.getByteString yLength return $ Signature x y instance Binary TxVersion where put = putTxVersion get = getTxVersion putTxVersion :: TxVersion -> Put putTxVersion (TxVersion v) = Put.putWord32le . fromIntegral $ v getTxVersion :: Get TxVersion getTxVersion = TxVersion . fromIntegral <$> Get.getWord32le defaultVersion :: TxVersion defaultVersion = TxVersion 1 defaultSequence :: Sequence defaultSequence = Sequence 0xffffffff instance Binary Sequence where put = putSequence get = getSequence putSequence :: Sequence -> Put putSequence (Sequence sequence') = Put.putWord32le sequence' getSequence :: Get Sequence getSequence = Sequence <$> Get.getWord32le defaultLockTime :: LockTime defaultLockTime = LockTime 0x00000000 instance Binary LockTime where put = putBlockLockTime get = getBlockLockTime putBlockLockTime :: LockTime -> Put putBlockLockTime (LockTime locktime') = Put.putWord32le locktime' getBlockLockTime :: Get LockTime getBlockLockTime = LockTime <$> Get.getWord32le putSighashAll :: Put putSighashAll = Put.putWord8 1 instance Arbitrary Transaction where arbitrary = do inputs' <- arbitrary outputs' <- arbitrary txVersion' <- arbitrary locktime' <- LockTime <$> arbitrary return Transaction { _inputs = inputs' , _outputs = outputs' , _txVersion = txVersion' , _locktime = locktime'} instance Arbitrary TxOutput where arbitrary = do value <- arbitrary script <- arbitrary return TxOutput { _value = value , _outputScript = script } instance Arbitrary TxInput where arbitrary = do utxo' <- arbitrary script <- arbitrary sequence' <- Sequence <$> arbitrary return TxInput { _utxo = utxo' , _signatureScript = script , _sequence = sequence'} instance Arbitrary TxVersion where arbitrary = TxVersion <$> choose (0, 0xffffffff) instance Arbitrary UTXO where arbitrary = do hash <- arbitrary index <- TxIndex <$> choose (0, 0xffffffff) return UTXO { _outTxHash = hash , _outIndex = index } instance Arbitrary Value where arbitrary = Satoshis <$> arbitrary `suchThat` (> 0)
clample/lamdabtc
backend/src/BitcoinCore/Transaction/Transactions.hs
bsd-3-clause
10,065
0
15
1,909
2,702
1,447
1,255
286
2
{-# LANGUAGE CPP #-} {-# LANGUAGE ConstraintKinds #-} {-# LANGUAGE DeriveDataTypeable #-} {-# LANGUAGE DeriveGeneric #-} {-# LANGUAGE FlexibleContexts #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE ScopedTypeVariables #-} {-# LANGUAGE TemplateHaskell #-} {-# LANGUAGE TupleSections #-} {-# LANGUAGE ViewPatterns #-} -- | Build the project. module Stack.Build (build ,withLoadPackage ,mkBaseConfigOpts ,queryBuildInfo ,splitObjsWarning ,CabalVersionException(..)) where import Control.Exception (Exception) import Control.Monad import Control.Monad.Catch (MonadMask, MonadMask) import Control.Monad.IO.Class import Control.Monad.Logger import Control.Monad.Reader (MonadReader, asks) import Control.Monad.Trans.Resource import Control.Monad.Trans.Unlift (MonadBaseUnlift) import Data.Aeson (Value (Object, Array), (.=), object) import Data.Function import qualified Data.HashMap.Strict as HM import Data.List ((\\)) import Data.List.Extra (groupSort) import Data.List.NonEmpty (NonEmpty(..)) import qualified Data.List.NonEmpty as NE import qualified Data.Map as Map import Data.Map.Strict (Map) import Data.Monoid import Data.Set (Set) import qualified Data.Set as Set import Data.Text (Text) import qualified Data.Text as T import Data.Text.Encoding (decodeUtf8) import qualified Data.Text.IO as TIO import Data.Text.Read (decimal) import Data.Typeable (Typeable) import qualified Data.Vector as V import qualified Data.Yaml as Yaml import Network.HTTP.Client.Conduit (HasHttpManager) import Path import Prelude hiding (FilePath, writeFile) import Stack.Build.ConstructPlan import Stack.Build.Execute import Stack.Build.Haddock import Stack.Build.Installed import Stack.Build.Source import Stack.Build.Target import Stack.Fetch as Fetch import Stack.GhcPkg import Stack.Package import Stack.Types.FlagName import Stack.Types.PackageIdentifier import Stack.Types.PackageName import Stack.Types.Version import Stack.Types.Config import Stack.Types.Build import Stack.Types.Package import Stack.Types.Internal #ifdef WINDOWS import Stack.Types.Compiler #endif import System.FileLock (FileLock, unlockFile) #ifdef WINDOWS import System.Win32.Console (setConsoleCP, setConsoleOutputCP, getConsoleCP, getConsoleOutputCP) import qualified Control.Monad.Catch as Catch #endif type M env m = (MonadIO m,MonadReader env m,HasHttpManager env,HasBuildConfig env,MonadLoggerIO m,MonadBaseUnlift IO m,MonadMask m,HasLogLevel env,HasEnvConfig env,HasTerminal env) -- | Build. -- -- If a buildLock is passed there is an important contract here. That lock must -- protect the snapshot, and it must be safe to unlock it if there are no further -- modifications to the snapshot to be performed by this build. build :: M env m => (Set (Path Abs File) -> IO ()) -- ^ callback after discovering all local files -> Maybe FileLock -> BuildOptsCLI -> m () build setLocalFiles mbuildLk boptsCli = fixCodePage $ do bopts <- asks (configBuild . getConfig) let profiling = boptsLibProfile bopts || boptsExeProfile bopts menv <- getMinimalEnvOverride (targets, mbp, locals, extraToBuild, sourceMap) <- loadSourceMap NeedTargets boptsCli -- Set local files, necessary for file watching stackYaml <- asks $ bcStackYaml . getBuildConfig liftIO $ setLocalFiles $ Set.insert stackYaml $ Set.unions $ map lpFiles locals (installedMap, globalDumpPkgs, snapshotDumpPkgs, localDumpPkgs) <- getInstalled menv GetInstalledOpts { getInstalledProfiling = profiling , getInstalledHaddock = shouldHaddockDeps bopts } sourceMap baseConfigOpts <- mkBaseConfigOpts boptsCli plan <- withLoadPackage menv $ \loadPackage -> constructPlan mbp baseConfigOpts locals extraToBuild localDumpPkgs loadPackage sourceMap installedMap -- If our work to do is all local, let someone else have a turn with the snapshot. -- They won't damage what's already in there. case (mbuildLk, allLocal plan) of -- NOTE: This policy is too conservative. In the future we should be able to -- schedule unlocking as an Action that happens after all non-local actions are -- complete. (Just lk,True) -> do $logDebug "All installs are local; releasing snapshot lock early." liftIO $ unlockFile lk _ -> return () checkCabalVersion warnAboutSplitObjs bopts warnIfExecutablesWithSameNameCouldBeOverwritten locals plan when (boptsPreFetch bopts) $ preFetch plan if boptsCLIDryrun boptsCli then printPlan plan else executePlan menv boptsCli baseConfigOpts locals globalDumpPkgs snapshotDumpPkgs localDumpPkgs installedMap targets plan -- | If all the tasks are local, they don't mutate anything outside of our local directory. allLocal :: Plan -> Bool allLocal = all (== Local) . map taskLocation . Map.elems . planTasks checkCabalVersion :: M env m => m () checkCabalVersion = do allowNewer <- asks (configAllowNewer . getConfig) cabalVer <- asks (envConfigCabalVersion . getEnvConfig) -- https://github.com/haskell/cabal/issues/2023 when (allowNewer && cabalVer < $(mkVersion "1.22")) $ throwM $ CabalVersionException $ "Error: --allow-newer requires at least Cabal version 1.22, but version " ++ versionString cabalVer ++ " was found." data CabalVersionException = CabalVersionException { unCabalVersionException :: String } deriving (Typeable) instance Show CabalVersionException where show = unCabalVersionException instance Exception CabalVersionException -- | See https://github.com/commercialhaskell/stack/issues/1198. warnIfExecutablesWithSameNameCouldBeOverwritten :: MonadLogger m => [LocalPackage] -> Plan -> m () warnIfExecutablesWithSameNameCouldBeOverwritten locals plan = do $logDebug "Checking if we are going to build multiple executables with the same name" forM_ (Map.toList warnings) $ \(exe,(toBuild,otherLocals)) -> do let exe_s | length toBuild > 1 = "several executables with the same name:" | otherwise = "executable" exesText pkgs = T.intercalate ", " ["'" <> packageNameText p <> ":" <> exe <> "'" | p <- pkgs] ($logWarn . T.unlines . concat) [ [ "Building " <> exe_s <> " " <> exesText toBuild <> "." ] , [ "Only one of them will be available via 'stack exec' or locally installed." | length toBuild > 1 ] , [ "Other executables with the same name might be overwritten: " <> exesText otherLocals <> "." | not (null otherLocals) ] ] where -- Cases of several local packages having executables with the same name. -- The Map entries have the following form: -- -- executable name: ( package names for executables that are being built -- , package names for other local packages that have an -- executable with the same name -- ) warnings :: Map Text ([PackageName],[PackageName]) warnings = Map.mapMaybe (\(pkgsToBuild,localPkgs) -> case (pkgsToBuild,NE.toList localPkgs \\ NE.toList pkgsToBuild) of (_ :| [],[]) -> -- We want to build the executable of single local package -- and there are no other local packages with an executable of -- the same name. Nothing to warn about, ignore. Nothing (_,otherLocals) -> -- We could be here for two reasons (or their combination): -- 1) We are building two or more executables with the same -- name that will end up overwriting each other. -- 2) In addition to the executable(s) that we want to build -- there are other local packages with an executable of the -- same name that might get overwritten. -- Both cases warrant a warning. Just (NE.toList pkgsToBuild,otherLocals)) (Map.intersectionWith (,) exesToBuild localExes) exesToBuild :: Map Text (NonEmpty PackageName) exesToBuild = collect [ (exe,pkgName) | (pkgName,task) <- Map.toList (planTasks plan) , isLocal task , exe <- (Set.toList . exeComponents . lpComponents . taskLP) task ] where isLocal Task{taskType = (TTLocal _)} = True isLocal _ = False taskLP Task{taskType = (TTLocal lp)} = lp taskLP _ = error "warnIfExecutablesWithSameNameCouldBeOverwritten/taskLP: task isn't local" localExes :: Map Text (NonEmpty PackageName) localExes = collect [ (exe,packageName pkg) | pkg <- map lpPackage locals , exe <- Set.toList (packageExes pkg) ] collect :: Ord k => [(k,v)] -> Map k (NonEmpty v) collect = Map.map NE.fromList . Map.fromDistinctAscList . groupSort warnAboutSplitObjs :: MonadLogger m => BuildOpts -> m () warnAboutSplitObjs bopts | boptsSplitObjs bopts = do $logWarn $ "Building with --split-objs is enabled. " <> T.pack splitObjsWarning warnAboutSplitObjs _ = return () splitObjsWarning :: String splitObjsWarning = unwords [ "Note that this feature is EXPERIMENTAL, and its behavior may be changed and improved." , "You will need to clean your workdirs before use. If you want to compile all dependencies" , "with split-objs, you will need to delete the snapshot (and all snapshots that could" , "reference that snapshot)." ] -- | Get the @BaseConfigOpts@ necessary for constructing configure options mkBaseConfigOpts :: (MonadIO m, MonadReader env m, HasEnvConfig env, MonadThrow m) => BuildOptsCLI -> m BaseConfigOpts mkBaseConfigOpts boptsCli = do bopts <- asks (configBuild . getConfig) snapDBPath <- packageDatabaseDeps localDBPath <- packageDatabaseLocal snapInstallRoot <- installationRootDeps localInstallRoot <- installationRootLocal packageExtraDBs <- packageDatabaseExtra return BaseConfigOpts { bcoSnapDB = snapDBPath , bcoLocalDB = localDBPath , bcoSnapInstallRoot = snapInstallRoot , bcoLocalInstallRoot = localInstallRoot , bcoBuildOpts = bopts , bcoBuildOptsCLI = boptsCli , bcoExtraDBs = packageExtraDBs } -- | Provide a function for loading package information from the package index withLoadPackage :: ( MonadIO m , HasHttpManager env , MonadReader env m , MonadBaseUnlift IO m , MonadMask m , MonadLogger m , HasEnvConfig env) => EnvOverride -> ((PackageName -> Version -> Map FlagName Bool -> [Text] -> IO Package) -> m a) -> m a withLoadPackage menv inner = do econfig <- asks getEnvConfig withCabalLoader menv $ \cabalLoader -> inner $ \name version flags ghcOptions -> do bs <- cabalLoader $ PackageIdentifier name version -- Intentionally ignore warnings, as it's not really -- appropriate to print a bunch of warnings out while -- resolving the package index. (_warnings,pkg) <- readPackageBS (depPackageConfig econfig flags ghcOptions) bs return pkg where -- | Package config to be used for dependencies depPackageConfig :: EnvConfig -> Map FlagName Bool -> [Text] -> PackageConfig depPackageConfig econfig flags ghcOptions = PackageConfig { packageConfigEnableTests = False , packageConfigEnableBenchmarks = False , packageConfigFlags = flags , packageConfigGhcOptions = ghcOptions , packageConfigCompilerVersion = envConfigCompilerVersion econfig , packageConfigPlatform = configPlatform (getConfig econfig) } -- | Set the code page for this process as necessary. Only applies to Windows. -- See: https://github.com/commercialhaskell/stack/issues/738 #ifdef WINDOWS fixCodePage :: M env m => m a -> m a fixCodePage inner = do mcp <- asks $ configModifyCodePage . getConfig ec <- asks getEnvConfig if mcp && getGhcVersion (envConfigCompilerVersion ec) < $(mkVersion "7.10.3") then fixCodePage' -- GHC >=7.10.3 doesn't need this code page hack. else inner where fixCodePage' = do origCPI <- liftIO getConsoleCP origCPO <- liftIO getConsoleOutputCP let setInput = origCPI /= expected setOutput = origCPO /= expected fixInput | setInput = Catch.bracket_ (liftIO $ do setConsoleCP expected) (liftIO $ setConsoleCP origCPI) | otherwise = id fixOutput | setInput = Catch.bracket_ (liftIO $ do setConsoleOutputCP expected) (liftIO $ setConsoleOutputCP origCPO) | otherwise = id case (setInput, setOutput) of (False, False) -> return () (True, True) -> warn "" (True, False) -> warn " input" (False, True) -> warn " output" fixInput $ fixOutput inner expected = 65001 -- UTF-8 warn typ = $logInfo $ T.concat [ "Setting" , typ , " codepage to UTF-8 (65001) to ensure correct output from GHC" ] #else fixCodePage :: a -> a fixCodePage = id #endif -- | Query information about the build and print the result to stdout in YAML format. queryBuildInfo :: M env m => [Text] -- ^ selectors -> m () queryBuildInfo selectors0 = rawBuildInfo >>= select id selectors0 >>= liftIO . TIO.putStrLn . decodeUtf8 . Yaml.encode where select _ [] value = return value select front (sel:sels) value = case value of Object o -> case HM.lookup sel o of Nothing -> err "Selector not found" Just value' -> cont value' Array v -> case decimal sel of Right (i, "") | i >= 0 && i < V.length v -> cont $ v V.! i | otherwise -> err "Index out of range" _ -> err "Encountered array and needed numeric selector" _ -> err $ "Cannot apply selector to " ++ show value where cont = select (front . (sel:)) sels err msg = error $ msg ++ ": " ++ show (front [sel]) -- | Get the raw build information object rawBuildInfo :: M env m => m Value rawBuildInfo = do (_, _mbp, locals, _extraToBuild, _sourceMap) <- loadSourceMap NeedTargets defaultBuildOptsCLI return $ object [ "locals" .= Object (HM.fromList $ map localToPair locals) ] where localToPair lp = (T.pack $ packageNameString $ packageName p, value) where p = lpPackage lp value = object [ "version" .= packageVersion p , "path" .= toFilePath (lpDir lp) ]
AndrewRademacher/stack
src/Stack/Build.hs
bsd-3-clause
16,147
0
20
5,022
3,238
1,725
1,513
264
6
module BrownPLT.JavaScript.Contracts.Interface ( module BrownPLT.JavaScript.Contracts.Types ) where import BrownPLT.JavaScript.Contracts.Types
brownplt/javascript-contracts
src/BrownPLT/JavaScript/Contracts/Interface.hs
bsd-3-clause
148
0
5
14
25
18
7
3
0
-- Module : Network.AWS.StorageGateway -- Copyright : (c) 2013-2014 Brendan Hay <[email protected]> -- License : This Source Code Form is subject to the terms of -- the Mozilla Public License, v. 2.0. -- A copy of the MPL can be found in the LICENSE file or -- you can obtain it at http://mozilla.org/MPL/2.0/. -- Maintainer : Brendan Hay <[email protected]> -- Stability : experimental -- Portability : non-portable (GHC extensions) -- -- Derived from AWS service descriptions, licensed under Apache 2.0. -- | AWS Storage Gateway is a service that connects an on-premises software -- appliance with cloud-based storage to provide seamless and secure integration -- between your on-premises IT environment and AWS's storage infrastructure. module Network.AWS.StorageGateway ( module Network.AWS.StorageGateway.ActivateGateway , module Network.AWS.StorageGateway.AddCache , module Network.AWS.StorageGateway.AddUploadBuffer , module Network.AWS.StorageGateway.AddWorkingStorage , module Network.AWS.StorageGateway.CancelArchival , module Network.AWS.StorageGateway.CancelRetrieval , module Network.AWS.StorageGateway.CreateCachediSCSIVolume , module Network.AWS.StorageGateway.CreateSnapshot , module Network.AWS.StorageGateway.CreateSnapshotFromVolumeRecoveryPoint , module Network.AWS.StorageGateway.CreateStorediSCSIVolume , module Network.AWS.StorageGateway.CreateTapes , module Network.AWS.StorageGateway.DeleteBandwidthRateLimit , module Network.AWS.StorageGateway.DeleteChapCredentials , module Network.AWS.StorageGateway.DeleteGateway , module Network.AWS.StorageGateway.DeleteSnapshotSchedule , module Network.AWS.StorageGateway.DeleteTape , module Network.AWS.StorageGateway.DeleteTapeArchive , module Network.AWS.StorageGateway.DeleteVolume , module Network.AWS.StorageGateway.DescribeBandwidthRateLimit , module Network.AWS.StorageGateway.DescribeCache , module Network.AWS.StorageGateway.DescribeCachediSCSIVolumes , module Network.AWS.StorageGateway.DescribeChapCredentials , module Network.AWS.StorageGateway.DescribeGatewayInformation , module Network.AWS.StorageGateway.DescribeMaintenanceStartTime , module Network.AWS.StorageGateway.DescribeSnapshotSchedule , module Network.AWS.StorageGateway.DescribeStorediSCSIVolumes , module Network.AWS.StorageGateway.DescribeTapeArchives , module Network.AWS.StorageGateway.DescribeTapeRecoveryPoints , module Network.AWS.StorageGateway.DescribeTapes , module Network.AWS.StorageGateway.DescribeUploadBuffer , module Network.AWS.StorageGateway.DescribeVTLDevices , module Network.AWS.StorageGateway.DescribeWorkingStorage , module Network.AWS.StorageGateway.DisableGateway , module Network.AWS.StorageGateway.ListGateways , module Network.AWS.StorageGateway.ListLocalDisks , module Network.AWS.StorageGateway.ListVolumeRecoveryPoints , module Network.AWS.StorageGateway.ListVolumes , module Network.AWS.StorageGateway.ResetCache , module Network.AWS.StorageGateway.RetrieveTapeArchive , module Network.AWS.StorageGateway.RetrieveTapeRecoveryPoint , module Network.AWS.StorageGateway.ShutdownGateway , module Network.AWS.StorageGateway.StartGateway , module Network.AWS.StorageGateway.Types , module Network.AWS.StorageGateway.UpdateBandwidthRateLimit , module Network.AWS.StorageGateway.UpdateChapCredentials , module Network.AWS.StorageGateway.UpdateGatewayInformation , module Network.AWS.StorageGateway.UpdateGatewaySoftwareNow , module Network.AWS.StorageGateway.UpdateMaintenanceStartTime , module Network.AWS.StorageGateway.UpdateSnapshotSchedule , module Network.AWS.StorageGateway.UpdateVTLDeviceType ) where import Network.AWS.StorageGateway.ActivateGateway import Network.AWS.StorageGateway.AddCache import Network.AWS.StorageGateway.AddUploadBuffer import Network.AWS.StorageGateway.AddWorkingStorage import Network.AWS.StorageGateway.CancelArchival import Network.AWS.StorageGateway.CancelRetrieval import Network.AWS.StorageGateway.CreateCachediSCSIVolume import Network.AWS.StorageGateway.CreateSnapshot import Network.AWS.StorageGateway.CreateSnapshotFromVolumeRecoveryPoint import Network.AWS.StorageGateway.CreateStorediSCSIVolume import Network.AWS.StorageGateway.CreateTapes import Network.AWS.StorageGateway.DeleteBandwidthRateLimit import Network.AWS.StorageGateway.DeleteChapCredentials import Network.AWS.StorageGateway.DeleteGateway import Network.AWS.StorageGateway.DeleteSnapshotSchedule import Network.AWS.StorageGateway.DeleteTape import Network.AWS.StorageGateway.DeleteTapeArchive import Network.AWS.StorageGateway.DeleteVolume import Network.AWS.StorageGateway.DescribeBandwidthRateLimit import Network.AWS.StorageGateway.DescribeCache import Network.AWS.StorageGateway.DescribeCachediSCSIVolumes import Network.AWS.StorageGateway.DescribeChapCredentials import Network.AWS.StorageGateway.DescribeGatewayInformation import Network.AWS.StorageGateway.DescribeMaintenanceStartTime import Network.AWS.StorageGateway.DescribeSnapshotSchedule import Network.AWS.StorageGateway.DescribeStorediSCSIVolumes import Network.AWS.StorageGateway.DescribeTapeArchives import Network.AWS.StorageGateway.DescribeTapeRecoveryPoints import Network.AWS.StorageGateway.DescribeTapes import Network.AWS.StorageGateway.DescribeUploadBuffer import Network.AWS.StorageGateway.DescribeVTLDevices import Network.AWS.StorageGateway.DescribeWorkingStorage import Network.AWS.StorageGateway.DisableGateway import Network.AWS.StorageGateway.ListGateways import Network.AWS.StorageGateway.ListLocalDisks import Network.AWS.StorageGateway.ListVolumeRecoveryPoints import Network.AWS.StorageGateway.ListVolumes import Network.AWS.StorageGateway.ResetCache import Network.AWS.StorageGateway.RetrieveTapeArchive import Network.AWS.StorageGateway.RetrieveTapeRecoveryPoint import Network.AWS.StorageGateway.ShutdownGateway import Network.AWS.StorageGateway.StartGateway import Network.AWS.StorageGateway.Types import Network.AWS.StorageGateway.UpdateBandwidthRateLimit import Network.AWS.StorageGateway.UpdateChapCredentials import Network.AWS.StorageGateway.UpdateGatewayInformation import Network.AWS.StorageGateway.UpdateGatewaySoftwareNow import Network.AWS.StorageGateway.UpdateMaintenanceStartTime import Network.AWS.StorageGateway.UpdateSnapshotSchedule import Network.AWS.StorageGateway.UpdateVTLDeviceType
romanb/amazonka
amazonka-storagegateway/gen/Network/AWS/StorageGateway.hs
mpl-2.0
6,542
0
5
630
773
570
203
101
0
{-# LANGUAGE TemplateHaskell #-} {-| Lenses for Ganeti config objects -} {- Copyright (C) 2014 Google Inc. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -} module Ganeti.Objects.Lens where import qualified Data.Set as Set import System.Time (ClockTime(..)) import Ganeti.Lens (makeCustomLenses, Lens') import Ganeti.Objects -- | Class of objects that have timestamps. class TimeStampObject a => TimeStampObjectL a where mTimeL :: Lens' a ClockTime -- | Class of objects that have an UUID. class UuidObject a => UuidObjectL a where uuidL :: Lens' a String -- | Class of object that have a serial number. class SerialNoObject a => SerialNoObjectL a where serialL :: Lens' a Int -- | Class of objects that have tags. class TagsObject a => TagsObjectL a where tagsL :: Lens' a (Set.Set String) $(makeCustomLenses ''AddressPool) $(makeCustomLenses ''Network) instance SerialNoObjectL Network where serialL = networkSerialL instance TagsObjectL Network where tagsL = networkTagsL instance UuidObjectL Network where uuidL = networkUuidL instance TimeStampObjectL Network where mTimeL = networkMtimeL $(makeCustomLenses ''PartialNic) $(makeCustomLenses ''Disk) $(makeCustomLenses ''Instance) instance TimeStampObjectL Instance where mTimeL = instMtimeL instance UuidObjectL Instance where uuidL = instUuidL instance SerialNoObjectL Instance where serialL = instSerialL instance TagsObjectL Instance where tagsL = instTagsL $(makeCustomLenses ''MinMaxISpecs) $(makeCustomLenses ''PartialIPolicy) $(makeCustomLenses ''FilledIPolicy) $(makeCustomLenses ''Node) instance TimeStampObjectL Node where mTimeL = nodeMtimeL instance UuidObjectL Node where uuidL = nodeUuidL instance SerialNoObjectL Node where serialL = nodeSerialL instance TagsObjectL Node where tagsL = nodeTagsL $(makeCustomLenses ''NodeGroup) instance TimeStampObjectL NodeGroup where mTimeL = groupMtimeL instance UuidObjectL NodeGroup where uuidL = groupUuidL instance SerialNoObjectL NodeGroup where serialL = groupSerialL instance TagsObjectL NodeGroup where tagsL = groupTagsL $(makeCustomLenses ''Cluster) instance TimeStampObjectL Cluster where mTimeL = clusterMtimeL instance UuidObjectL Cluster where uuidL = clusterUuidL instance SerialNoObjectL Cluster where serialL = clusterSerialL instance TagsObjectL Cluster where tagsL = clusterTagsL $(makeCustomLenses ''ConfigData) instance SerialNoObjectL ConfigData where serialL = configSerialL instance TimeStampObjectL ConfigData where mTimeL = configMtimeL
apyrgio/ganeti
src/Ganeti/Objects/Lens.hs
bsd-2-clause
3,779
0
10
587
596
304
292
70
0
module List2 where import FFI import Prelude hiding (take) main :: Fay () main = putStrLn (showList (take 5 (let ns = 1 : map' (foo 123) ns in ns))) foo :: Double -> Double -> Double foo x y = x * y / 2 take :: Int -> [a] -> [a] take 0 _ = [] take n (x:xs) = x : take (n - 1) xs map' :: (a -> b) -> [a] -> [b] map' f [] = [] map' f (x:xs) = f x : map' f xs showList :: [Double] -> String showList = ffi "JSON.stringify(%1)"
fpco/fay
tests/List2.hs
bsd-3-clause
460
0
18
138
261
138
123
15
1
{-| Module : Idris.ProofSearch Description : Searches current context for proofs' Copyright : License : BSD3 Maintainer : The Idris Community. -} {-# LANGUAGE PatternGuards #-} module Idris.ProofSearch( trivial , trivialHoles , proofSearch , resolveTC ) where import Idris.Core.Elaborate hiding (Tactic(..)) import Idris.Core.TT import Idris.Core.Unify import Idris.Core.Evaluate import Idris.Core.CaseTree import Idris.Core.Typecheck import Idris.AbsSyntax import Idris.Delaborate import Idris.Error import Control.Applicative ((<$>)) import Control.Monad import Control.Monad.State.Strict import qualified Data.Set as S import Data.List import Debug.Trace -- Pass in a term elaborator to avoid a cyclic dependency with ElabTerm trivial :: (PTerm -> ElabD ()) -> IState -> ElabD () trivial = trivialHoles [] [] trivialHoles :: [Name] -> -- user visible names, when working -- in interactive mode [(Name, Int)] -> (PTerm -> ElabD ()) -> IState -> ElabD () trivialHoles psnames ok elab ist = try' (do elab (PApp (fileFC "prf") (PRef (fileFC "prf") [] eqCon) [pimp (sUN "A") Placeholder False, pimp (sUN "x") Placeholder False]) return ()) (do env <- get_env g <- goal tryAll env return ()) True where tryAll [] = fail "No trivial solution" tryAll ((x, b):xs) = do -- if type of x has any holes in it, move on hs <- get_holes let badhs = hs -- filter (flip notElem holesOK) hs g <- goal -- anywhere but the top is okay for a hole, if holesOK set if -- all (\n -> not (n `elem` badhs)) (freeNames (binderTy b)) (holesOK hs (binderTy b) && (null psnames || x `elem` psnames)) then try' (elab (PRef (fileFC "prf") [] x)) (tryAll xs) True else tryAll xs holesOK hs ap@(App _ _ _) | (P _ n _, args) <- unApply ap = holeArgsOK hs n 0 args holesOK hs (App _ f a) = holesOK hs f && holesOK hs a holesOK hs (P _ n _) = not (n `elem` hs) holesOK hs (Bind n b sc) = holesOK hs (binderTy b) && holesOK hs sc holesOK hs _ = True holeArgsOK hs n p [] = True holeArgsOK hs n p (a : as) | (n, p) `elem` ok = holeArgsOK hs n (p + 1) as | otherwise = holesOK hs a && holeArgsOK hs n (p + 1) as trivialTCs :: [(Name, Int)] -> (PTerm -> ElabD ()) -> IState -> ElabD () trivialTCs ok elab ist = try' (do elab (PApp (fileFC "prf") (PRef (fileFC "prf") [] eqCon) [pimp (sUN "A") Placeholder False, pimp (sUN "x") Placeholder False]) return ()) (do env <- get_env g <- goal tryAll env return ()) True where tryAll [] = fail "No trivial solution" tryAll ((x, b):xs) = do -- if type of x has any holes in it, move on hs <- get_holes let badhs = hs -- filter (flip notElem holesOK) hs g <- goal env <- get_env -- anywhere but the top is okay for a hole, if holesOK set if -- all (\n -> not (n `elem` badhs)) (freeNames (binderTy b)) (holesOK hs (binderTy b) && tcArg env (binderTy b)) then try' (elab (PRef (fileFC "prf") [] x)) (tryAll xs) True else tryAll xs tcArg env ty | (P _ n _, args) <- unApply (getRetTy (normalise (tt_ctxt ist) env ty)) = case lookupCtxtExact n (idris_classes ist) of Just _ -> True _ -> False | otherwise = False holesOK hs ap@(App _ _ _) | (P _ n _, args) <- unApply ap = holeArgsOK hs n 0 args holesOK hs (App _ f a) = holesOK hs f && holesOK hs a holesOK hs (P _ n _) = not (n `elem` hs) holesOK hs (Bind n b sc) = holesOK hs (binderTy b) && holesOK hs sc holesOK hs _ = True holeArgsOK hs n p [] = True holeArgsOK hs n p (a : as) | (n, p) `elem` ok = holeArgsOK hs n (p + 1) as | otherwise = holesOK hs a && holeArgsOK hs n (p + 1) as cantSolveGoal :: ElabD a cantSolveGoal = do g <- goal env <- get_env lift $ tfail $ CantSolveGoal g (map (\(n,b) -> (n, binderTy b)) env) proofSearch :: Bool -- ^ recursive search (False for 'refine') -> Bool -- ^ invoked from a tactic proof. If so, making new metavariables is meaningless, and there should be an error reported instead. -> Bool -- ^ ambiguity ok -> Bool -- ^ defer on failure -> Int -- ^ maximum depth -> (PTerm -> ElabD ()) -> Maybe Name -> Name -> [Name] -> [Name] -> IState -> ElabD () proofSearch False fromProver ambigok deferonfail depth elab _ nroot psnames [fn] ist = do -- get all possible versions of the name, take the first one that -- works let all_imps = lookupCtxtName fn (idris_implicits ist) tryAllFns all_imps where -- if nothing worked, make a new metavariable tryAllFns [] | fromProver = cantSolveGoal tryAllFns [] = do attack; defer [] nroot; solve tryAllFns (f : fs) = try' (tryFn f) (tryAllFns fs) True tryFn (f, args) = do let imps = map isImp args ps <- get_probs hs <- get_holes args <- map snd <$> try' (apply (Var f) imps) (match_apply (Var f) imps) True ps' <- get_probs -- when (length ps < length ps') $ fail "Can't apply constructor" -- Make metavariables for new holes hs' <- get_holes ptm <- get_term if fromProver then cantSolveGoal else do mapM_ (\ h -> do focus h attack; defer [] nroot; solve) (hs' \\ hs) -- (filter (\ (x, y) -> not x) (zip (map fst imps) args)) solve isImp (PImp p _ _ _ _) = (True, p) isImp arg = (True, priority arg) -- try to get all of them by unification proofSearch rec fromProver ambigok deferonfail maxDepth elab fn nroot psnames hints ist = do compute ty <- goal hs <- get_holes env <- get_env tm <- get_term argsok <- conArgsOK ty if ambigok || argsok then case lookupCtxt nroot (idris_tyinfodata ist) of [TISolution ts] -> findInferredTy ts _ -> if ambigok then psRec rec maxDepth [] S.empty -- postpone if it fails early in elaboration else handleError cantsolve (psRec rec maxDepth [] S.empty) (autoArg (sUN "auto")) else autoArg (sUN "auto") -- not enough info in the type yet where findInferredTy (t : _) = elab (delab ist (toUN t)) cantsolve (InternalMsg _) = True cantsolve (CantSolveGoal _ _) = True cantsolve (IncompleteTerm _) = True cantsolve (At _ e) = cantsolve e cantsolve (Elaborating _ _ _ e) = cantsolve e cantsolve (ElaboratingArg _ _ _ e) = cantsolve e cantsolve err = False conArgsOK ty = let (f, as) = unApply ty in case f of P _ n _ -> let autohints = case lookupCtxtExact n (idris_autohints ist) of Nothing -> [] Just hs -> hs in case lookupCtxtExact n (idris_datatypes ist) of Just t -> do rs <- mapM (conReady as) (autohints ++ con_names t) return (and rs) Nothing -> -- local variable, go for it return True TType _ -> return True _ -> typeNotSearchable ty conReady :: [Term] -> Name -> ElabD Bool conReady as n = case lookupTyExact n (tt_ctxt ist) of Just ty -> do let (_, cs) = unApply (getRetTy ty) -- if any metavariables in 'as' correspond to -- a constructor form in 'cs', then we're not -- ready to run auto yet. Otherwise, go for it hs <- get_holes return $ and (map (notHole hs) (zip as cs)) Nothing -> fail "Can't happen" -- if n is a metavariable, and c is a constructor form, we're not ready -- to run yet notHole hs (P _ n _, c) | (P _ cn _, _) <- unApply c, n `elem` hs && isConName cn (tt_ctxt ist) = False | Constant _ <- c = not (n `elem` hs) -- if fa is a metavariable applied to anything, we're not ready to run yet. notHole hs (fa, c) | (P _ fn _, args@(_:_)) <- unApply fa = fn `notElem` hs notHole _ _ = True inHS hs (P _ n _) = n `elem` hs isHS _ _ = False toUN t@(P nt (MN i n) ty) | ('_':xs) <- str n = t | otherwise = P nt (UN n) ty toUN (App s f a) = App s (toUN f) (toUN a) toUN t = t -- psRec counts depth and the local variable applications we're under -- (so we don't try a pointless application of something to itself, -- which obviously won't work anyway but might lead us on a wild -- goose chase...) -- Also keep track of the types we've proved so far in this branch -- (if we get back to one we've been to before, we're just in a cycle and -- that's no use) psRec :: Bool -> Int -> [Name] -> S.Set Type -> ElabD () psRec _ 0 locs tys | fromProver = cantSolveGoal psRec rec 0 locs tys = do attack; defer [] nroot; solve --fail "Maximum depth reached" psRec False d locs tys = tryCons d locs tys hints psRec True d locs tys = do compute ty <- goal when (S.member ty tys) $ fail "Been here before" let tys' = S.insert ty tys try' (try' (trivialHoles psnames [] elab ist) (resolveTC False False 20 ty nroot elab ist) True) (try' (try' (resolveByCon (d - 1) locs tys') (resolveByLocals (d - 1) locs tys') True) -- if all else fails, make a new metavariable (if fromProver then fail "cantSolveGoal" else do attack; defer [] nroot; solve) True) True -- get recursive function name. Only user given names make sense. getFn d (Just f) | d < maxDepth-1 && usersname f = [f] | otherwise = [] getFn d _ = [] usersname (UN _) = True usersname (NS n _) = usersname n usersname _ = False resolveByCon d locs tys = do t <- goal let (f, _) = unApply t case f of P _ n _ -> do let autohints = case lookupCtxtExact n (idris_autohints ist) of Nothing -> [] Just hs -> hs case lookupCtxtExact n (idris_datatypes ist) of Just t -> do let others = hints ++ con_names t ++ autohints tryCons d locs tys (others ++ getFn d fn) Nothing -> typeNotSearchable t _ -> typeNotSearchable t -- if there are local variables which have a function type, try -- applying them too resolveByLocals d locs tys = do env <- get_env tryLocals d locs tys env tryLocals d locs tys [] = fail "Locals failed" tryLocals d locs tys ((x, t) : xs) | x `elem` locs || x `notElem` psnames = tryLocals d locs tys xs | otherwise = try' (tryLocal d (x : locs) tys x t) (tryLocals d locs tys xs) True tryCons d locs tys [] = fail "Constructors failed" tryCons d locs tys (c : cs) = try' (tryCon d locs tys c) (tryCons d locs tys cs) True tryLocal d locs tys n t = do let a = getPArity (delab ist (binderTy t)) tryLocalArg d locs tys n a tryLocalArg d locs tys n 0 = elab (PRef (fileFC "prf") [] n) tryLocalArg d locs tys n i = simple_app False (tryLocalArg d locs tys n (i - 1)) (psRec True d locs tys) "proof search local apply" -- Like interface resolution, but searching with constructors tryCon d locs tys n = do ty <- goal let imps = case lookupCtxtExact n (idris_implicits ist) of Nothing -> [] Just args -> map isImp args ps <- get_probs hs <- get_holes args <- map snd <$> try' (apply (Var n) imps) (match_apply (Var n) imps) True ps' <- get_probs hs' <- get_holes when (length ps < length ps') $ fail "Can't apply constructor" let newhs = filter (\ (x, y) -> not x) (zip (map fst imps) args) mapM_ (\ (_, h) -> do focus h aty <- goal psRec True d locs tys) newhs solve isImp (PImp p _ _ _ _) = (True, p) isImp arg = (False, priority arg) typeNotSearchable ty = lift $ tfail $ FancyMsg $ [TextPart "Attempted to find an element of type", TermPart ty, TextPart "using proof search, but proof search only works on datatypes with constructors."] ++ case ty of (Bind _ (Pi _ _ _) _) -> [TextPart "In particular, function types are not supported."] _ -> [] -- | Resolve interfaces. This will only pick up 'normal' -- implementations, never named implementations (which is enforced by -- 'findInstances'). resolveTC :: Bool -- ^ using default Int -> Bool -- ^ allow open implementations -> Int -- ^ depth -> Term -- ^ top level goal, for error messages -> Name -- ^ top level function name, to prevent loops -> (PTerm -> ElabD ()) -- ^ top level elaborator -> IState -> ElabD () resolveTC def openOK depth top fn elab ist = do hs <- get_holes resTC' [] def openOK hs depth top fn elab ist resTC' tcs def openOK topholes 0 topg fn elab ist = fail "Can't resolve interface" resTC' tcs def openOK topholes 1 topg fn elab ist = try' (trivial elab ist) (resolveTC def False 0 topg fn elab ist) True resTC' tcs defaultOn openOK topholes depth topg fn elab ist = do compute if openOK then try' (resolveOpen (idris_openimpls ist)) resolveNormal True else resolveNormal where -- try all the Open implementations first resolveOpen open = do t <- goal blunderbuss t depth [] open resolveNormal = do -- Resolution can proceed only if there is something concrete in the -- determining argument positions. Keep track of the holes in the -- non-determining position, because it's okay for 'trivial' to solve -- those holes and no others. g <- goal let (argsok, okholePos) = case tcArgsOK g topholes of Nothing -> (False, []) Just hs -> (True, hs) env <- get_env probs <- get_probs if not argsok -- && not mvok) then lift $ tfail $ CantResolve True topg (probErr probs) else do ptm <- get_term ulog <- getUnifyLog hs <- get_holes env <- get_env t <- goal let (tc, ttypes) = unApply (getRetTy t) let okholes = case tc of P _ n _ -> zip (repeat n) okholePos _ -> [] traceWhen ulog ("Resolving class " ++ show g ++ "\nin" ++ show env ++ "\n" ++ show okholes) $ try' (trivialTCs okholes elab ist) (do addDefault t tc ttypes let stk = map fst (filter snd $ elab_stack ist) let insts = idris_openimpls ist ++ findInstances ist t blunderbuss t depth stk (stk ++ insts)) True -- returns Just hs if okay, where hs are holes which are okay in the -- goal, or Nothing if not okay to proceed tcArgsOK ty hs | (P _ nc _, as) <- unApply (getRetTy ty), nc == numclass && defaultOn = Just [] tcArgsOK ty hs -- if any determining arguments are metavariables, postpone = let (f, as) = unApply (getRetTy ty) in case f of P _ cn _ -> case lookupCtxtExact cn (idris_classes ist) of Just ci -> tcDetArgsOK 0 (class_determiners ci) hs as Nothing -> if any (isMeta hs) as then Nothing else Just [] _ -> if any (isMeta hs) as then Nothing else Just [] -- return the list of argument positions which can safely be a hole -- or Nothing if one of the determining arguments is a hole tcDetArgsOK i ds hs (x : xs) | i `elem` ds = if isMeta hs x then Nothing else tcDetArgsOK (i + 1) ds hs xs | otherwise = do rs <- tcDetArgsOK (i + 1) ds hs xs case x of P _ n _ -> Just (i : rs) _ -> Just rs tcDetArgsOK _ _ _ [] = Just [] probErr [] = Msg "" probErr ((_,_,_,_,err,_,_) : _) = err isMeta :: [Name] -> Term -> Bool isMeta ns (P _ n _) = n `elem` ns isMeta _ _ = False notHole hs (P _ n _, c) | (P _ cn _, _) <- unApply (getRetTy c), n `elem` hs && isConName cn (tt_ctxt ist) = False | Constant _ <- c = not (n `elem` hs) notHole _ _ = True numclass = sNS (sUN "Num") ["Interfaces","Prelude"] addDefault t num@(P _ nc _) [P Bound a _] | nc == numclass && defaultOn = do focus a fill (RConstant (AType (ATInt ITBig))) -- default Integer solve addDefault t f as | all boundVar as = return () -- True -- fail $ "Can't resolve " ++ show t addDefault t f a = return () -- trace (show t) $ return () boundVar (P Bound _ _) = True boundVar _ = False blunderbuss t d stk [] = do ps <- get_probs lift $ tfail $ CantResolve False topg (probErr ps) blunderbuss t d stk (n:ns) | n /= fn -- && (n `elem` stk) = tryCatch (resolve n d) (\e -> case e of CantResolve True _ _ -> lift $ tfail e _ -> blunderbuss t d stk ns) | otherwise = blunderbuss t d stk ns introImps = do g <- goal case g of (Bind _ (Pi _ _ _) sc) -> do attack; intro Nothing num <- introImps return (num + 1) _ -> return 0 solven n = replicateM_ n solve resolve n depth | depth == 0 = fail "Can't resolve interface" | otherwise = do lams <- introImps t <- goal let (tc, ttypes) = trace (show t) $ unApply (getRetTy t) -- if (all boundVar ttypes) then resolveTC (depth - 1) fn insts ist -- else do -- if there's a hole in the goal, don't even try let imps = case lookupCtxtName n (idris_implicits ist) of [] -> [] [args] -> map isImp (snd args) -- won't be overloaded! xs -> error "The impossible happened - overloading is not expected here!" ps <- get_probs tm <- get_term args <- map snd <$> apply (Var n) imps solven lams -- close any implicit lambdas we introduced ps' <- get_probs when (length ps < length ps' || unrecoverable ps') $ fail "Can't apply interface" -- traceWhen (all boundVar ttypes) ("Progress: " ++ show t ++ " with " ++ show n) $ mapM_ (\ (_,n) -> do focus n t' <- goal let (tc', ttype) = unApply (getRetTy t') let got = fst (unApply (getRetTy t)) let depth' = if tc' `elem` tcs then depth - 1 else depth resTC' (got : tcs) defaultOn openOK topholes depth' topg fn elab ist) (filter (\ (x, y) -> not x) (zip (map fst imps) args)) -- if there's any arguments left, we've failed to resolve hs <- get_holes ulog <- getUnifyLog solve traceWhen ulog ("Got " ++ show n) $ return () where isImp (PImp p _ _ _ _) = (True, p) isImp arg = (False, priority arg) -- | Find the names of implementations that have been designeated for -- searching (i.e. non-named implementations or implementations from Elab scripts) findInstances :: IState -> Term -> [Name] findInstances ist t | (P _ n _, _) <- unApply (getRetTy t) = case lookupCtxt n (idris_classes ist) of [CI _ _ _ _ _ ins _] -> [n | (n, True) <- ins, accessible n] _ -> [] | otherwise = [] where accessible n = case lookupDefAccExact n False (tt_ctxt ist) of Just (_, Hidden) -> False Just (_, Private) -> False _ -> True
tpsinnem/Idris-dev
src/Idris/ProofSearch.hs
bsd-3-clause
23,017
160
22
9,737
6,142
3,222
2,920
422
39
module DeadRecursiveBoxed where import Prelude topEntity :: Bool -> Bool topEntity y = f ((\x -> x) . (\y -> y)) y f x y = fst (x, f x y) y
christiaanb/clash-compiler
tests/shouldwork/BoxedFunctions/DeadRecursiveBoxed.hs
bsd-2-clause
143
0
9
35
78
43
35
5
1
{-# LANGUAGE OverloadedStrings, FlexibleInstances #-} {-# OPTIONS_GHC -fno-warn-orphans #-} module Documentation.Haddock.ParserSpec (main, spec) where import Data.String import qualified Documentation.Haddock.Parser as Parse import Documentation.Haddock.Types import Documentation.Haddock.Doc (docAppend) import Test.Hspec import Test.QuickCheck infixr 6 <> (<>) :: Doc id -> Doc id -> Doc id (<>) = docAppend type Doc id = DocH () id instance IsString (Doc String) where fromString = DocString instance IsString a => IsString (Maybe a) where fromString = Just . fromString parseParas :: String -> MetaDoc () String parseParas = overDoc Parse.toRegular . Parse.parseParas parseString :: String -> Doc String parseString = Parse.toRegular . Parse.parseString hyperlink :: String -> Maybe String -> Doc String hyperlink url = DocHyperlink . Hyperlink url main :: IO () main = hspec spec spec :: Spec spec = do describe "parseString" $ do let infix 1 `shouldParseTo` shouldParseTo :: String -> Doc String -> Expectation shouldParseTo input ast = parseString input `shouldBe` ast it "is total" $ do property $ \xs -> (length . show . parseString) xs `shouldSatisfy` (> 0) context "when parsing text" $ do it "can handle unicode" $ do "灼眼のシャナ" `shouldParseTo` "灼眼のシャナ" it "accepts numeric character references" $ do "foo b&#97;r b&#97;z &#955;" `shouldParseTo` "foo bar baz λ" it "accepts hexadecimal character references" $ do "&#x65;" `shouldParseTo` "e" it "allows to backslash-escape characters" $ do property $ \x -> ['\\', x] `shouldParseTo` DocString [x] context "when parsing strings contaning numeric character references" $ do it "will implicitly convert digits to characters" $ do "&#65;&#65;&#65;&#65;" `shouldParseTo` "AAAA" "&#28796;&#30524;&#12398;&#12471;&#12515;&#12490;" `shouldParseTo` "灼眼のシャナ" it "will implicitly convert hex encoded characters" $ do "&#x65;&#x65;&#x65;&#x65;" `shouldParseTo` "eeee" context "when parsing identifiers" $ do it "parses identifiers enclosed within single ticks" $ do "'foo'" `shouldParseTo` DocIdentifier "foo" it "parses identifiers enclosed within backticks" $ do "`foo`" `shouldParseTo` DocIdentifier "foo" it "parses a word with an one of the delimiters in it as DocString" $ do "don't" `shouldParseTo` "don't" it "doesn't pass pairs of delimiters with spaces between them" $ do "hel'lo w'orld" `shouldParseTo` "hel'lo w'orld" it "don't use apostrophe's in the wrong place's" $ do " don't use apostrophe's in the wrong place's" `shouldParseTo` "don't use apostrophe's in the wrong place's" it "doesn't parse empty identifiers" $ do "``" `shouldParseTo` "``" it "can parse infix identifiers" $ do "``infix``" `shouldParseTo` "`" <> DocIdentifier "infix" <> "`" context "when parsing URLs" $ do it "parses a URL" $ do "<http://example.com/>" `shouldParseTo` hyperlink "http://example.com/" Nothing it "accepts an optional label" $ do "<http://example.com/ some link>" `shouldParseTo` hyperlink "http://example.com/" "some link" it "does not accept newlines in label" $ do "<foo bar\nbaz>" `shouldParseTo` "<foo bar\nbaz>" -- new behaviour test, this will be now consistent with other markup it "allows us to escape > inside the URL" $ do "<http://examp\\>le.com>" `shouldParseTo` hyperlink "http://examp>le.com" Nothing "<http://exa\\>mp\\>le.com>" `shouldParseTo` hyperlink "http://exa>mp>le.com" Nothing -- Likewise in label "<http://example.com f\\>oo>" `shouldParseTo` hyperlink "http://example.com" "f>oo" it "parses inline URLs" $ do "foo <http://example.com/> bar" `shouldParseTo` "foo " <> hyperlink "http://example.com/" Nothing <> " bar" it "doesn't allow for multi-line link tags" $ do "<ba\nz aar>" `shouldParseTo` "<ba\nz aar>" context "when parsing markdown links" $ do it "parses a simple link" $ do "[some label](url)" `shouldParseTo` hyperlink "url" "some label" it "allows whitespace between label and URL" $ do "[some label] \t (url)" `shouldParseTo` hyperlink "url" "some label" it "allows newlines in label" $ do "[some\n\nlabel](url)" `shouldParseTo` hyperlink "url" "some\n\nlabel" it "allows escaping in label" $ do "[some\\] label](url)" `shouldParseTo` hyperlink "url" "some] label" it "strips leading and trailing whitespace from label" $ do "[ some label ](url)" `shouldParseTo` hyperlink "url" "some label" it "rejects whitespace in URL" $ do "[some label]( url)" `shouldParseTo` "[some label]( url)" context "when URL is on a separate line" $ do it "allows URL to be on a separate line" $ do "[some label]\n(url)" `shouldParseTo` hyperlink "url" "some label" it "allows leading whitespace" $ do "[some label]\n \t (url)" `shouldParseTo` hyperlink "url" "some label" it "rejects additional newlines" $ do "[some label]\n\n(url)" `shouldParseTo` "[some label]\n\n(url)" context "when autolinking URLs" $ do it "autolinks HTTP URLs" $ do "http://example.com/" `shouldParseTo` hyperlink "http://example.com/" Nothing it "autolinks HTTPS URLs" $ do "https://www.example.com/" `shouldParseTo` hyperlink "https://www.example.com/" Nothing it "autolinks FTP URLs" $ do "ftp://example.com/" `shouldParseTo` hyperlink "ftp://example.com/" Nothing it "does not include a trailing comma" $ do "http://example.com/, Some other sentence." `shouldParseTo` hyperlink "http://example.com/" Nothing <> ", Some other sentence." it "does not include a trailing dot" $ do "http://example.com/. Some other sentence." `shouldParseTo` hyperlink "http://example.com/" Nothing <> ". Some other sentence." it "does not include a trailing exclamation mark" $ do "http://example.com/! Some other sentence." `shouldParseTo` hyperlink "http://example.com/" Nothing <> "! Some other sentence." it "does not include a trailing question mark" $ do "http://example.com/? Some other sentence." `shouldParseTo` hyperlink "http://example.com/" Nothing <> "? Some other sentence." it "autolinks URLs occuring mid-sentence with multiple ‘/’s" $ do "foo https://example.com/example bar" `shouldParseTo` "foo " <> hyperlink "https://example.com/example" Nothing <> " bar" context "when parsing images" $ do let image :: String -> Maybe String -> Doc String image uri = DocPic . Picture uri it "accepts markdown syntax for images" $ do "![label](url)" `shouldParseTo` image "url" "label" it "accepts Unicode" $ do "![灼眼のシャナ](url)" `shouldParseTo` image "url" "灼眼のシャナ" it "supports deprecated picture syntax" $ do "<<baz>>" `shouldParseTo` image "baz" Nothing it "supports title for deprecated picture syntax" $ do "<<b a z>>" `shouldParseTo` image "b" "a z" context "when parsing anchors" $ do it "parses a single word anchor" $ do "#foo#" `shouldParseTo` DocAName "foo" it "parses a multi word anchor" $ do "#foo bar#" `shouldParseTo` DocAName "foo bar" it "parses a unicode anchor" $ do "#灼眼のシャナ#" `shouldParseTo` DocAName "灼眼のシャナ" it "does not accept newlines in anchors" $ do "#foo\nbar#" `shouldParseTo` "#foo\nbar#" it "accepts anchors mid-paragraph" $ do "Hello #someAnchor# world!" `shouldParseTo` "Hello " <> DocAName "someAnchor" <> " world!" it "does not accept empty anchors" $ do "##" `shouldParseTo` "##" context "when parsing emphasised text" $ do it "emphasises a word on its own" $ do "/foo/" `shouldParseTo` DocEmphasis "foo" it "emphasises inline correctly" $ do "foo /bar/ baz" `shouldParseTo` "foo " <> DocEmphasis "bar" <> " baz" it "emphasises unicode" $ do "/灼眼のシャナ/" `shouldParseTo` DocEmphasis "灼眼のシャナ" it "does not emphasise multi-line strings" $ do " /foo\nbar/" `shouldParseTo` "/foo\nbar/" it "does not emphasise the empty string" $ do "//" `shouldParseTo` "//" it "parses escaped slashes literally" $ do "/foo\\/bar/" `shouldParseTo` DocEmphasis "foo/bar" it "recognizes other markup constructs within emphasised text" $ do "/foo @bar@ baz/" `shouldParseTo` DocEmphasis ("foo " <> DocMonospaced "bar" <> " baz") it "allows other markup inside of emphasis" $ do "/__inner bold__/" `shouldParseTo` DocEmphasis (DocBold "inner bold") it "doesn't mangle inner markup unicode" $ do "/__灼眼のシャナ &#65;__/" `shouldParseTo` DocEmphasis (DocBold "灼眼のシャナ A") it "properly converts HTML escape sequences" $ do "/&#65;&#65;&#65;&#65;/" `shouldParseTo` DocEmphasis "AAAA" it "allows to escape the emphasis delimiter inside of emphasis" $ do "/empha\\/sis/" `shouldParseTo` DocEmphasis "empha/sis" context "when parsing monospaced text" $ do it "parses simple monospaced text" $ do "@foo@" `shouldParseTo` DocMonospaced "foo" it "parses inline monospaced text" $ do "foo @bar@ baz" `shouldParseTo` "foo " <> DocMonospaced "bar" <> " baz" it "allows to escape @" $ do "@foo \\@ bar@" `shouldParseTo` DocMonospaced "foo @ bar" it "accepts unicode" $ do "@foo 灼眼のシャナ bar@" `shouldParseTo` DocMonospaced "foo 灼眼のシャナ bar" it "accepts other markup in monospaced text" $ do "@/foo/@" `shouldParseTo` DocMonospaced (DocEmphasis "foo") it "requires the closing @" $ do "@foo /bar/ baz" `shouldParseTo` "@foo " <> DocEmphasis "bar" <> " baz" context "when parsing bold strings" $ do it "allows for a bold string on its own" $ do "__bold string__" `shouldParseTo` DocBold "bold string" it "bolds inline correctly" $ do "hello __everyone__ there" `shouldParseTo` "hello " <> DocBold "everyone" <> " there" it "bolds unicode" $ do "__灼眼のシャナ__" `shouldParseTo` DocBold "灼眼のシャナ" it "does not do __multi-line\\n bold__" $ do " __multi-line\n bold__" `shouldParseTo` "__multi-line\n bold__" it "allows other markup inside of bold" $ do "__/inner emphasis/__" `shouldParseTo` (DocBold $ DocEmphasis "inner emphasis") it "doesn't mangle inner markup unicode" $ do "__/灼眼のシャナ &#65;/__" `shouldParseTo` (DocBold $ DocEmphasis "灼眼のシャナ A") it "properly converts HTML escape sequences" $ do "__&#65;&#65;&#65;&#65;__" `shouldParseTo` DocBold "AAAA" it "allows to escape the bold delimiter inside of bold" $ do "__bo\\__ld__" `shouldParseTo` DocBold "bo__ld" it "doesn't allow for empty bold" $ do "____" `shouldParseTo` "____" context "when parsing module strings" $ do it "should parse a module on its own" $ do "\"Module\"" `shouldParseTo` DocModule "Module" it "should parse a module inline" $ do "This is a \"Module\"." `shouldParseTo` "This is a " <> DocModule "Module" <> "." it "can accept a simple module name" $ do "\"Hello\"" `shouldParseTo` DocModule "Hello" it "can accept a module name with dots" $ do "\"Hello.World\"" `shouldParseTo` DocModule "Hello.World" it "can accept a module name with unicode" $ do "\"Hello.Worldλ\"" `shouldParseTo` DocModule "Hello.Worldλ" it "parses a module name with a trailing dot as regular quoted string" $ do "\"Hello.\"" `shouldParseTo` "\"Hello.\"" it "parses a module name with a space as regular quoted string" $ do "\"Hello World\"" `shouldParseTo` "\"Hello World\"" it "parses a module name with invalid characters as regular quoted string" $ do "\"Hello&[{}(=*)+]!\"" `shouldParseTo` "\"Hello&[{}(=*)+]!\"" it "accepts a module name with unicode" $ do "\"Foo.Barλ\"" `shouldParseTo` DocModule "Foo.Barλ" it "treats empty module name as regular double quotes" $ do "\"\"" `shouldParseTo` "\"\"" it "accepts anchor reference syntax as DocModule" $ do "\"Foo#bar\"" `shouldParseTo` DocModule "Foo#bar" it "accepts old anchor reference syntax as DocModule" $ do "\"Foo\\#bar\"" `shouldParseTo` DocModule "Foo\\#bar" describe "parseParas" $ do let infix 1 `shouldParseTo` shouldParseTo :: String -> Doc String -> Expectation shouldParseTo input ast = _doc (parseParas input) `shouldBe` ast it "is total" $ do property $ \xs -> (length . show . parseParas) xs `shouldSatisfy` (> 0) context "when parsing @since" $ do it "adds specified version to the result" $ do parseParas "@since 0.5.0" `shouldBe` MetaDoc { _meta = Meta { _version = Just [0,5,0] } , _doc = DocEmpty } it "ignores trailing whitespace" $ do parseParas "@since 0.5.0 \t " `shouldBe` MetaDoc { _meta = Meta { _version = Just [0,5,0] } , _doc = DocEmpty } it "does not allow trailing input" $ do parseParas "@since 0.5.0 foo" `shouldBe` MetaDoc { _meta = Meta { _version = Nothing } , _doc = DocParagraph "@since 0.5.0 foo" } context "when given multiple times" $ do it "gives last occurrence precedence" $ do (parseParas . unlines) [ "@since 0.5.0" , "@since 0.6.0" , "@since 0.7.0" ] `shouldBe` MetaDoc { _meta = Meta { _version = Just [0,7,0] } , _doc = DocEmpty } context "when parsing text paragraphs" $ do let filterSpecial = filter (`notElem` (".(=#-[*`\v\f\n\t\r\\\"'_/@<> " :: String)) it "parses an empty paragraph" $ do "" `shouldParseTo` DocEmpty it "parses a simple text paragraph" $ do "foo bar baz" `shouldParseTo` DocParagraph "foo bar baz" it "accepts markup in text paragraphs" $ do "foo /bar/ baz" `shouldParseTo` DocParagraph ("foo " <> DocEmphasis "bar" <> " baz") it "preserve all regular characters" $ do property $ \xs -> let input = filterSpecial xs in (not . null) input ==> input `shouldParseTo` DocParagraph (DocString input) it "separates paragraphs by empty lines" $ do unlines [ "foo" , " \t " , "bar" ] `shouldParseTo` DocParagraph "foo" <> DocParagraph "bar" context "when a pragraph only contains monospaced text" $ do it "turns it into a code block" $ do "@foo@" `shouldParseTo` DocCodeBlock "foo" context "when a paragraph starts with a markdown link" $ do it "correctly parses it as a text paragraph (not a definition list)" $ do "[label](url)" `shouldParseTo` DocParagraph (hyperlink "url" "label") it "can be followed by an other paragraph" $ do "[label](url)\n\nfoobar" `shouldParseTo` DocParagraph (hyperlink "url" "label") <> DocParagraph "foobar" context "when paragraph contains additional text" $ do it "accepts more text after the link" $ do "[label](url) foo bar baz" `shouldParseTo` DocParagraph (hyperlink "url" "label" <> " foo bar baz") it "accepts a newline right after the markdown link" $ do "[label](url)\nfoo bar baz" `shouldParseTo` DocParagraph (hyperlink "url" "label" <> " foo bar baz") it "can be followed by an other paragraph" $ do "[label](url)foo\n\nbar" `shouldParseTo` DocParagraph (hyperlink "url" "label" <> "foo") <> DocParagraph "bar" context "when parsing birdtracks" $ do it "parses them as a code block" $ do unlines [ ">foo" , ">bar" , ">baz" ] `shouldParseTo` DocCodeBlock "foo\nbar\nbaz" it "ignores leading whitespace" $ do unlines [ " >foo" , " \t >bar" , " >baz" ] `shouldParseTo` DocCodeBlock "foo\nbar\nbaz" it "strips one leading space from each line of the block" $ do unlines [ "> foo" , "> bar" , "> baz" ] `shouldParseTo` DocCodeBlock "foo\n bar\nbaz" it "ignores empty lines when stripping spaces" $ do unlines [ "> foo" , ">" , "> bar" ] `shouldParseTo` DocCodeBlock "foo\n\nbar" context "when any non-empty line does not start with a space" $ do it "does not strip any spaces" $ do unlines [ ">foo" , "> bar" ] `shouldParseTo` DocCodeBlock "foo\n bar" it "ignores nested markup" $ do unlines [ ">/foo/" ] `shouldParseTo` DocCodeBlock "/foo/" it "treats them as regular text inside text paragraphs" $ do unlines [ "foo" , ">bar" ] `shouldParseTo` DocParagraph "foo\n>bar" context "when parsing code blocks" $ do it "accepts a simple code block" $ do unlines [ "@" , "foo" , "bar" , "baz" , "@" ] `shouldParseTo` DocCodeBlock "foo\nbar\nbaz\n" it "ignores trailing whitespace after the opening @" $ do unlines [ "@ " , "foo" , "@" ] `shouldParseTo` DocCodeBlock "foo\n" it "rejects code blocks that are not closed" $ do unlines [ "@" , "foo" ] `shouldParseTo` DocParagraph "@\nfoo" it "accepts nested markup" $ do unlines [ "@" , "/foo/" , "@" ] `shouldParseTo` DocCodeBlock (DocEmphasis "foo" <> "\n") it "allows to escape the @" $ do unlines [ "@" , "foo" , "\\@" , "bar" , "@" ] `shouldParseTo` DocCodeBlock "foo\n@\nbar\n" it "accepts horizontal space before the @" $ do unlines [ " @" , "foo" , "" , "bar" , "@" ] `shouldParseTo` DocCodeBlock "foo\n\nbar\n" it "strips a leading space from a @ block if present" $ do unlines [ " @" , " hello" , " world" , " @" ] `shouldParseTo` DocCodeBlock "hello\nworld\n" unlines [ " @" , " hello" , "" , " world" , " @" ] `shouldParseTo` DocCodeBlock "hello\n\nworld\n" it "only drops whitespace if there's some before closing @" $ do unlines [ "@" , " Formatting" , " matters." , "@" ] `shouldParseTo` DocCodeBlock " Formatting\n matters.\n" it "accepts unicode" $ do "@foo 灼眼のシャナ bar@" `shouldParseTo` DocCodeBlock "foo 灼眼のシャナ bar" it "requires the closing @" $ do "@foo /bar/ baz" `shouldParseTo` DocParagraph ("@foo " <> DocEmphasis "bar" <> " baz") context "when parsing examples" $ do it "parses a simple example" $ do ">>> foo" `shouldParseTo` DocExamples [Example "foo" []] it "parses an example with result" $ do unlines [ ">>> foo" , "bar" , "baz" ] `shouldParseTo` DocExamples [Example "foo" ["bar", "baz"]] it "parses consecutive examples" $ do unlines [ ">>> fib 5" , "5" , ">>> fib 10" , "55" ] `shouldParseTo` DocExamples [ Example "fib 5" ["5"] , Example "fib 10" ["55"] ] it ("requires an example to be separated" ++ " from a previous paragraph by an empty line") $ do "foobar\n\n>>> fib 10\n55" `shouldParseTo` DocParagraph "foobar" <> DocExamples [Example "fib 10" ["55"]] it "parses bird-tracks inside of paragraphs as plain strings" $ do let xs = "foo\n>>> bar" xs `shouldParseTo` DocParagraph (DocString xs) it "skips empty lines in front of an example" $ do "\n \n\n>>> foo" `shouldParseTo` DocExamples [Example "foo" []] it "terminates example on empty line" $ do unlines [ ">>> foo" , "bar" , " " , "baz" ] `shouldParseTo` DocExamples [Example "foo" ["bar"]] <> DocParagraph "baz" it "parses a <BLANKLINE> result as an empty result" $ do unlines [ ">>> foo" , "bar" , "<BLANKLINE>" , "baz" ] `shouldParseTo` DocExamples [Example "foo" ["bar", "", "baz"]] it "accepts unicode in examples" $ do ">>> 灼眼\nシャナ" `shouldParseTo` DocExamples [Example "灼眼" ["シャナ"]] context "when prompt is prefixed by whitespace" $ do it "strips the exact same amount of whitespace from result lines" $ do unlines [ " >>> foo" , " bar" , " baz" ] `shouldParseTo` DocExamples [Example "foo" ["bar", "baz"]] it "preserves additional whitespace" $ do unlines [ " >>> foo" , " bar" ] `shouldParseTo` DocExamples [Example "foo" [" bar"]] it "keeps original if stripping is not possible" $ do unlines [ " >>> foo" , " bar" ] `shouldParseTo` DocExamples [Example "foo" [" bar"]] context "when parsing paragraphs nested in lists" $ do it "can nest the same type of list" $ do "* foo\n\n * bar" `shouldParseTo` DocUnorderedList [ DocParagraph "foo" <> DocUnorderedList [DocParagraph "bar"]] it "can nest another type of list inside" $ do "* foo\n\n 1. bar" `shouldParseTo` DocUnorderedList [ DocParagraph "foo" <> DocOrderedList [DocParagraph "bar"]] it "can nest a code block inside" $ do "* foo\n\n @foo bar baz@" `shouldParseTo` DocUnorderedList [ DocParagraph "foo" <> DocCodeBlock "foo bar baz"] "* foo\n\n @\n foo bar baz\n @" `shouldParseTo` DocUnorderedList [ DocParagraph "foo" <> DocCodeBlock "foo bar baz\n"] it "can nest more than one level" $ do "* foo\n\n * bar\n\n * baz\n qux" `shouldParseTo` DocUnorderedList [ DocParagraph "foo" <> DocUnorderedList [ DocParagraph "bar" <> DocUnorderedList [DocParagraph "baz\nqux"] ] ] it "won't fail on not fully indented paragraph" $ do "* foo\n\n * bar\n\n * qux\nquux" `shouldParseTo` DocUnorderedList [ DocParagraph "foo" <> DocUnorderedList [ DocParagraph "bar" ] , DocParagraph "qux\nquux"] it "can nest definition lists" $ do "[a]: foo\n\n [b]: bar\n\n [c]: baz\n qux" `shouldParseTo` DocDefList [ ("a", "foo" <> DocDefList [ ("b", "bar" <> DocDefList [("c", "baz\nqux")]) ]) ] it "can come back to top level with a different list" $ do "* foo\n\n * bar\n\n1. baz" `shouldParseTo` DocUnorderedList [ DocParagraph "foo" <> DocUnorderedList [ DocParagraph "bar" ] ] <> DocOrderedList [ DocParagraph "baz" ] it "definition lists can come back to top level with a different list" $ do "[foo]: foov\n\n [bar]: barv\n\n1. baz" `shouldParseTo` DocDefList [ ("foo", "foov" <> DocDefList [ ("bar", "barv") ]) ] <> DocOrderedList [ DocParagraph "baz" ] it "list order is preserved in presence of nesting + extra text" $ do "1. Foo\n\n > Some code\n\n2. Bar\n\nSome text" `shouldParseTo` DocOrderedList [ DocParagraph "Foo" <> DocCodeBlock "Some code" , DocParagraph "Bar" ] <> DocParagraph (DocString "Some text") "1. Foo\n\n2. Bar\n\nSome text" `shouldParseTo` DocOrderedList [ DocParagraph "Foo" , DocParagraph "Bar" ] <> DocParagraph (DocString "Some text") context "when parsing properties" $ do it "can parse a single property" $ do "prop> 23 == 23" `shouldParseTo` DocProperty "23 == 23" it "can parse multiple subsequent properties" $ do unlines [ "prop> 23 == 23" , "prop> 42 == 42" ] `shouldParseTo` DocProperty "23 == 23" <> DocProperty "42 == 42" it "accepts unicode in properties" $ do "prop> 灼眼のシャナ ≡ 愛" `shouldParseTo` DocProperty "灼眼のシャナ ≡ 愛" it "can deal with whitespace before and after the prop> prompt" $ do " prop> xs == (reverse $ reverse xs) " `shouldParseTo` DocProperty "xs == (reverse $ reverse xs)" context "when parsing unordered lists" $ do it "parses a simple list" $ do unlines [ " * one" , " * two" , " * three" ] `shouldParseTo` DocUnorderedList [ DocParagraph "one" , DocParagraph "two" , DocParagraph "three" ] it "ignores empty lines between list items" $ do unlines [ "* one" , "" , "* two" ] `shouldParseTo` DocUnorderedList [ DocParagraph "one" , DocParagraph "two" ] it "accepts an empty list item" $ do "*" `shouldParseTo` DocUnorderedList [DocParagraph DocEmpty] it "accepts multi-line list items" $ do unlines [ "* point one" , " more one" , "* point two" , "more two" ] `shouldParseTo` DocUnorderedList [ DocParagraph "point one\n more one" , DocParagraph "point two\nmore two" ] it "accepts markup in list items" $ do "* /foo/" `shouldParseTo` DocUnorderedList [DocParagraph (DocEmphasis "foo")] it "requires empty lines between list and other paragraphs" $ do unlines [ "foo" , "" , "* bar" , "" , "baz" ] `shouldParseTo` DocParagraph "foo" <> DocUnorderedList [DocParagraph "bar"] <> DocParagraph "baz" context "when parsing ordered lists" $ do it "parses a simple list" $ do unlines [ " 1. one" , " (1) two" , " 3. three" ] `shouldParseTo` DocOrderedList [ DocParagraph "one" , DocParagraph "two" , DocParagraph "three" ] it "ignores empty lines between list items" $ do unlines [ "1. one" , "" , "2. two" ] `shouldParseTo` DocOrderedList [ DocParagraph "one" , DocParagraph "two" ] it "accepts an empty list item" $ do "1." `shouldParseTo` DocOrderedList [DocParagraph DocEmpty] it "accepts multi-line list items" $ do unlines [ "1. point one" , " more one" , "1. point two" , "more two" ] `shouldParseTo` DocOrderedList [ DocParagraph "point one\n more one" , DocParagraph "point two\nmore two" ] it "accepts markup in list items" $ do "1. /foo/" `shouldParseTo` DocOrderedList [DocParagraph (DocEmphasis "foo")] it "requires empty lines between list and other paragraphs" $ do unlines [ "foo" , "" , "1. bar" , "" , "baz" ] `shouldParseTo` DocParagraph "foo" <> DocOrderedList [DocParagraph "bar"] <> DocParagraph "baz" context "when parsing definition lists" $ do it "parses a simple list" $ do unlines [ " [foo]: one" , " [bar]: two" , " [baz]: three" ] `shouldParseTo` DocDefList [ ("foo", "one") , ("bar", "two") , ("baz", "three") ] it "ignores empty lines between list items" $ do unlines [ "[foo]: one" , "" , "[bar]: two" ] `shouldParseTo` DocDefList [ ("foo", "one") , ("bar", "two") ] it "accepts an empty list item" $ do "[foo]:" `shouldParseTo` DocDefList [("foo", DocEmpty)] it "accepts multi-line list items" $ do unlines [ "[foo]: point one" , " more one" , "[bar]: point two" , "more two" ] `shouldParseTo` DocDefList [ ("foo", "point one\n more one") , ("bar", "point two\nmore two") ] it "accepts markup in list items" $ do "[foo]: /foo/" `shouldParseTo` DocDefList [("foo", DocEmphasis "foo")] it "accepts markup for the label" $ do "[/foo/]: bar" `shouldParseTo` DocDefList [(DocEmphasis "foo", "bar")] it "requires empty lines between list and other paragraphs" $ do unlines [ "foo" , "" , "[foo]: bar" , "" , "baz" ] `shouldParseTo` DocParagraph "foo" <> DocDefList [("foo", "bar")] <> DocParagraph "baz" it "dose not require the colon (deprecated - this will be removed in a future release)" $ do unlines [ " [foo] one" , " [bar] two" , " [baz] three" ] `shouldParseTo` DocDefList [ ("foo", "one") , ("bar", "two") , ("baz", "three") ] context "when parsing consecutive paragraphs" $ do it "will not capture irrelevant consecutive lists" $ do unlines [ " * bullet" , "" , "" , " - different bullet" , "" , "" , " (1) ordered" , " " , " 2. different bullet" , " " , " [cat]: kitten" , " " , " [pineapple]: fruit" ] `shouldParseTo` DocUnorderedList [ DocParagraph "bullet" , DocParagraph "different bullet"] <> DocOrderedList [ DocParagraph "ordered" , DocParagraph "different bullet" ] <> DocDefList [ ("cat", "kitten") , ("pineapple", "fruit") ] context "when parsing function documentation headers" $ do it "can parse a simple header" $ do "= Header 1\nHello." `shouldParseTo` (DocHeader (Header 1 "Header 1")) <> DocParagraph "Hello." it "allow consecutive headers" $ do "= Header 1\n== Header 2" `shouldParseTo` DocHeader (Header 1 "Header 1") <> DocHeader (Header 2 "Header 2") it "accepts markup in the header" $ do "= /Header/ __1__\nFoo" `shouldParseTo` DocHeader (Header 1 (DocEmphasis "Header" <> " " <> DocBold "1")) <> DocParagraph "Foo"
DavidAlphaFox/ghc
utils/haddock/haddock-library/test/Documentation/Haddock/ParserSpec.hs
bsd-3-clause
33,032
0
76
11,371
6,446
3,142
3,304
713
1
-- Copyright (c) 2014-present, Facebook, Inc. -- All rights reserved. -- -- This source code is distributed under the terms of a BSD license, -- found in the LICENSE file. An additional grant of patent rights can -- be found in the PATENTS file. {-# LANGUAGE CPP #-} {-# LANGUAGE BangPatterns #-} {-# LANGUAGE ConstraintKinds #-} {-# LANGUAGE ExistentialQuantification #-} {-# LANGUAGE FlexibleContexts #-} {-# LANGUAGE GADTs #-} {-# LANGUAGE MultiWayIf #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE PatternGuards #-} {-# LANGUAGE ScopedTypeVariables #-} -- | The implementation of the 'Haxl' monad. module Haxl.Core.Monad ( -- * The monad GenHaxl (..), runHaxl, env, -- * Env Env(..), caches, initEnvWithData, initEnv, emptyEnv, -- * Exceptions throw, catch, catchIf, try, tryToHaxlException, -- * Data fetching and caching dataFetch, uncachedRequest, cacheRequest, cacheResult, cachedComputation, dumpCacheAsHaskell, -- * Unsafe operations unsafeLiftIO, unsafeToHaxlException, ) where import Haxl.Core.Types import Haxl.Core.Show1 import Haxl.Core.StateStore import Haxl.Core.Exception import Haxl.Core.RequestStore import Haxl.Core.Util import Haxl.Core.DataCache as DataCache import qualified Data.Text as Text import Control.Exception (Exception(..), SomeException) #if __GLASGOW_HASKELL__ >= 708 import Control.Exception (SomeAsyncException(..)) #endif #if __GLASGOW_HASKELL__ >= 710 import Control.Exception (AllocationLimitExceeded(..)) #endif import Control.Monad import qualified Control.Exception as Exception import Control.Applicative hiding (Const) import Control.DeepSeq import GHC.Exts (IsString(..)) #if __GLASGOW_HASKELL__ < 706 import Prelude hiding (catch) #endif import Data.IORef import Data.List import Data.Monoid import Data.Time import qualified Data.HashMap.Strict as HashMap import Text.Printf import Text.PrettyPrint hiding ((<>)) import Control.Arrow (left) #ifdef EVENTLOG import Control.Exception (bracket_) import Debug.Trace (traceEventIO) #endif -- ----------------------------------------------------------------------------- -- The environment -- | The data we carry around in the Haxl monad. data Env u = Env { cacheRef :: IORef (DataCache ResultVar) -- cached data fetches , memoRef :: IORef (DataCache (MemoVar u)) -- memoized computations , flags :: Flags , userEnv :: u , statsRef :: IORef Stats , states :: StateStore -- ^ Data sources and other components can store their state in -- here. Items in this store must be instances of 'StateKey'. } type Caches u = (IORef (DataCache ResultVar), IORef (DataCache (MemoVar u))) caches :: Env u -> Caches u caches env = (cacheRef env, memoRef env) -- | Initialize an environment with a 'StateStore', an input map, a -- preexisting 'DataCache', and a seed for the random number generator. initEnvWithData :: StateStore -> u -> Caches u -> IO (Env u) initEnvWithData states e (cref, mref) = do sref <- newIORef emptyStats return Env { cacheRef = cref , memoRef = mref , flags = defaultFlags , userEnv = e , states = states , statsRef = sref } -- | Initializes an environment with 'DataStates' and an input map. initEnv :: StateStore -> u -> IO (Env u) initEnv states e = do cref <- newIORef DataCache.empty mref <- newIORef DataCache.empty initEnvWithData states e (cref,mref) -- | A new, empty environment. emptyEnv :: u -> IO (Env u) emptyEnv = initEnv stateEmpty -- ----------------------------------------------------------------------------- -- | The Haxl monad, which does several things: -- -- * It is a reader monad for 'Env' and 'IORef' 'RequestStore', The -- latter is the current batch of unsubmitted data fetch requests. -- -- * It is a concurrency, or resumption, monad. A computation may run -- partially and return 'Blocked', in which case the framework should -- perform the outstanding requests in the 'RequestStore', and then -- resume the computation. -- -- * The Applicative combinator '<*>' explores /both/ branches in the -- event that the left branch is 'Blocked', so that we can collect -- multiple requests and submit them as a batch. -- -- * It contains IO, so that we can perform real data fetching. -- newtype GenHaxl u a = GenHaxl { unHaxl :: Env u -> IORef (RequestStore u) -> IO (Result u a) } -- | The result of a computation is either 'Done' with a value, 'Throw' -- with an exception, or 'Blocked' on the result of a data fetch with -- a continuation. data Result u a = Done a | Throw SomeException | Blocked (GenHaxl u a) instance (Show a) => Show (Result u a) where show (Done a) = printf "Done(%s)" $ show a show (Throw e) = printf "Throw(%s)" $ show e show Blocked{} = "Blocked" instance Monad (GenHaxl u) where return a = GenHaxl $ \_env _ref -> return (Done a) GenHaxl m >>= k = GenHaxl $ \env ref -> do e <- m env ref case e of Done a -> unHaxl (k a) env ref Throw e -> return (Throw e) Blocked cont -> return (Blocked (cont >>= k)) instance Functor (GenHaxl u) where fmap f m = pure f <*> m instance Applicative (GenHaxl u) where pure = return GenHaxl f <*> GenHaxl a = GenHaxl $ \env ref -> do r <- f env ref case r of Throw e -> return (Throw e) Done f' -> do ra <- a env ref case ra of Done a' -> return (Done (f' a')) Throw e -> return (Throw e) Blocked a' -> return (Blocked (f' <$> a')) Blocked f' -> do ra <- a env ref -- left is blocked, explore the right case ra of Done a' -> return (Blocked (f' <*> return a')) Throw e -> return (Blocked (f' <*> throw e)) Blocked a' -> return (Blocked (f' <*> a')) -- | Runs a 'Haxl' computation in an 'Env'. runHaxl :: Env u -> GenHaxl u a -> IO a #ifdef EVENTLOG runHaxl env h = do let go !n env (GenHaxl haxl) = do traceEventIO "START computation" ref <- newIORef noRequests e <- haxl env ref traceEventIO "STOP computation" case e of Done a -> return a Throw e -> Exception.throw e Blocked cont -> do bs <- readIORef ref writeIORef ref noRequests -- Note [RoundId] traceEventIO "START performFetches" n' <- performFetches n env bs traceEventIO "STOP performFetches" go n' env cont traceEventIO "START runHaxl" r <- go 0 env h traceEventIO "STOP runHaxl" return r #else runHaxl env (GenHaxl haxl) = do ref <- newIORef noRequests e <- haxl env ref case e of Done a -> return a Throw e -> Exception.throw e Blocked cont -> do bs <- readIORef ref writeIORef ref noRequests -- Note [RoundId] void (performFetches 0 env bs) runHaxl env cont #endif -- | Extracts data from the 'Env'. env :: (Env u -> a) -> GenHaxl u a env f = GenHaxl $ \env _ref -> return (Done (f env)) -- ----------------------------------------------------------------------------- -- Exceptions -- | Throw an exception in the Haxl monad throw :: (Exception e) => e -> GenHaxl u a throw e = GenHaxl $ \_env _ref -> raise e raise :: (Exception e) => e -> IO (Result u a) raise = return . Throw . toException -- | Catch an exception in the Haxl monad catch :: Exception e => GenHaxl u a -> (e -> GenHaxl u a) -> GenHaxl u a catch (GenHaxl m) h = GenHaxl $ \env ref -> do r <- m env ref case r of Done a -> return (Done a) Throw e | Just e' <- fromException e -> unHaxl (h e') env ref | otherwise -> return (Throw e) Blocked k -> return (Blocked (catch k h)) -- | Catch exceptions that satisfy a predicate catchIf :: Exception e => (e -> Bool) -> GenHaxl u a -> (e -> GenHaxl u a) -> GenHaxl u a catchIf cond haxl handler = catch haxl $ \e -> if cond e then handler e else throw e -- | Returns @'Left' e@ if the computation throws an exception @e@, or -- @'Right' a@ if it returns a result @a@. try :: Exception e => GenHaxl u a -> GenHaxl u (Either e a) try haxl = (Right <$> haxl) `catch` (return . Left) -- ----------------------------------------------------------------------------- -- Unsafe operations -- | Under ordinary circumstances this is unnecessary; users of the Haxl -- monad should generally /not/ perform arbitrary IO. unsafeLiftIO :: IO a -> GenHaxl u a unsafeLiftIO m = GenHaxl $ \_env _ref -> Done <$> m -- | Convert exceptions in the underlying IO monad to exceptions in -- the Haxl monad. This is morally unsafe, because you could then -- catch those exceptions in Haxl and observe the underlying execution -- order. Not to be exposed to user code. unsafeToHaxlException :: GenHaxl u a -> GenHaxl u a unsafeToHaxlException (GenHaxl m) = GenHaxl $ \env ref -> do r <- m env ref `Exception.catch` \e -> return (Throw e) case r of Blocked c -> return (Blocked (unsafeToHaxlException c)) other -> return other -- | Like 'try', but lifts all exceptions into the 'HaxlException' -- hierarchy. Uses 'unsafeToHaxlException' internally. Typically -- this is used at the top level of a Haxl computation, to ensure that -- all exceptions are caught. tryToHaxlException :: GenHaxl u a -> GenHaxl u (Either HaxlException a) tryToHaxlException h = left asHaxlException <$> try (unsafeToHaxlException h) -- ----------------------------------------------------------------------------- -- Data fetching and caching -- | Possible responses when checking the cache. data CacheResult a -- | The request hadn't been seen until now. = Uncached (ResultVar a) -- | The request has been seen before, but its result has not yet been -- fetched. | CachedNotFetched (ResultVar a) -- | The request has been seen before, and its result has already been -- fetched. | Cached (Either SomeException a) -- | Checks the data cache for the result of a request. cached :: (Request r a) => Env u -> r a -> IO (CacheResult a) cached env req = do cache <- readIORef (cacheRef env) let do_fetch = do rvar <- newEmptyResult writeIORef (cacheRef env) $! DataCache.insert req rvar cache return (Uncached rvar) case DataCache.lookup req cache of Nothing -> do_fetch Just rvar -> do mb <- tryReadResult rvar case mb of Nothing -> return (CachedNotFetched rvar) -- Use the cached result, even if it was an error. Just r -> do ifTrace (flags env) 3 $ putStrLn $ case r of Left _ -> "Cached error: " ++ show req Right _ -> "Cached request: " ++ show req return (Cached r) -- | Performs actual fetching of data for a 'Request' from a 'DataSource'. dataFetch :: (DataSource u r, Request r a) => r a -> GenHaxl u a dataFetch req = GenHaxl $ \env ref -> do -- First, check the cache res <- cached env req case res of -- Not seen before: add the request to the RequestStore, so it -- will be fetched in the next round. Uncached rvar -> do modifyIORef' ref $ \bs -> addRequest (BlockedFetch req rvar) bs return $ Blocked (continueFetch req rvar) -- Seen before but not fetched yet. We're blocked, but we don't have -- to add the request to the RequestStore. CachedNotFetched rvar -> return $ Blocked (continueFetch req rvar) -- Cached: either a result, or an exception Cached (Left ex) -> return (Throw ex) Cached (Right a) -> return (Done a) -- | A data request that is not cached. This is not what you want for -- normal read requests, because then multiple identical requests may -- return different results, and this invalidates some of the -- properties that we expect Haxl computations to respect: that data -- fetches can be aribtrarily reordered, and identical requests can be -- commoned up, for example. -- -- 'uncachedRequest' is useful for performing writes, provided those -- are done in a safe way - that is, not mixed with reads that might -- conflict in the same Haxl computation. -- uncachedRequest :: (DataSource u r, Request r a) => r a -> GenHaxl u a uncachedRequest req = GenHaxl $ \_env ref -> do rvar <- newEmptyResult modifyIORef' ref $ \bs -> addRequest (BlockedFetch req rvar) bs return $ Blocked (continueFetch req rvar) continueFetch :: (DataSource u r, Request r a, Show a) => r a -> ResultVar a -> GenHaxl u a continueFetch req rvar = GenHaxl $ \_env _ref -> do m <- tryReadResult rvar case m of Nothing -> raise . DataSourceError $ textShow req <> " did not set contents of result var" Just r -> done r -- | Transparently provides caching. Useful for datasources that can -- return immediately, but also caches values. Exceptions thrown by -- the IO operation (except for asynchronous exceptions) are -- propagated into the Haxl monad and can be caught by 'catch' and -- 'try'. cacheResult :: (Request r a) => r a -> IO a -> GenHaxl u a cacheResult req val = GenHaxl $ \env _ref -> do cachedResult <- cached env req case cachedResult of Uncached rvar -> do result <- Exception.try val putResult rvar result case result of Left e -> do rethrowAsyncExceptions e; done result _other -> done result Cached result -> done result CachedNotFetched _ -> corruptCache where corruptCache = raise . DataSourceError $ Text.concat [ textShow req , " has a corrupted cache value: these requests are meant to" , " return immediately without an intermediate value. Either" , " the cache was updated incorrectly, or you're calling" , " cacheResult on a query that involves a blocking fetch." ] -- We must be careful about turning IO monad exceptions into Haxl -- exceptions. An IO monad exception will normally propagate right -- out of runHaxl and terminate the whole computation, whereas a Haxl -- exception can get dropped on the floor, if it is on the right of -- <*> and the left side also throws, for example. So turning an IO -- monad exception into a Haxl exception is a dangerous thing to do. -- In particular, we never want to do it for an asynchronous exception -- (AllocationLimitExceeded, ThreadKilled, etc.), because these are -- supposed to unconditionally terminate the computation. -- -- There are three places where we take an arbitrary IO monad exception and -- turn it into a Haxl exception: -- -- * wrapFetchInCatch. Here we want to propagate a failure of the -- data source to the callers of the data source, but if the -- failure came from elsewhere (an asynchronous exception), then we -- should just propagate it -- -- * cacheResult (cache the results of IO operations): again, -- failures of the IO operation should be visible to the caller as -- a Haxl exception, but we exclude asynchronous exceptions from -- this. -- * unsafeToHaxlException: assume the caller knows what they're -- doing, and just wrap all exceptions. -- rethrowAsyncExceptions :: SomeException -> IO () rethrowAsyncExceptions e #if __GLASGOW_HASKELL__ >= 708 | Just SomeAsyncException{} <- fromException e = Exception.throw e #endif #if __GLASGOW_HASKELL__ >= 710 | Just AllocationLimitExceeded{} <- fromException e = Exception.throw e -- AllocationLimitExceeded is not a child of SomeAsyncException, -- but it should be. #endif | otherwise = return () -- | Inserts a request/result pair into the cache. Throws an exception -- if the request has already been issued, either via 'dataFetch' or -- 'cacheRequest'. -- -- This can be used to pre-populate the cache when running tests, to -- avoid going to the actual data source and ensure that results are -- deterministic. -- cacheRequest :: (Request req a) => req a -> Either SomeException a -> GenHaxl u () cacheRequest request result = GenHaxl $ \env _ref -> do res <- cached env request case res of Uncached rvar -> do -- request was not in the cache: insert the result and continue putResult rvar result return $ Done () -- It is an error if the request is already in the cache. We can't test -- whether the cached result is the same without adding an Eq constraint, -- and we don't necessarily have Eq for all results. _other -> raise $ DataSourceError "cacheRequest: request is already in the cache" instance IsString a => IsString (GenHaxl u a) where fromString s = return (fromString s) -- | Issues a batch of fetches in a 'RequestStore'. After -- 'performFetches', all the requests in the 'RequestStore' are -- complete, and all of the 'ResultVar's are full. performFetches :: forall u. Int -> Env u -> RequestStore u -> IO Int performFetches n env reqs = do let f = flags env sref = statsRef env jobs = contents reqs !n' = n + length jobs t0 <- getCurrentTime let roundstats = [ (dataSourceName (getReq reqs), length reqs) | BlockedFetches reqs <- jobs ] where getReq :: [BlockedFetch r] -> r a getReq = undefined ifTrace f 1 $ printf "Batch data fetch (%s)\n" $ intercalate (", "::String) $ map (\(name,num) -> printf "%d %s" num (Text.unpack name)) roundstats ifTrace f 3 $ forM_ jobs $ \(BlockedFetches reqs) -> forM_ reqs $ \(BlockedFetch r _) -> putStrLn (show1 r) let applyFetch (i, BlockedFetches (reqs :: [BlockedFetch r])) = case stateGet (states env) of Nothing -> return (SyncFetch (mapM_ (setError (const e)) reqs)) where req :: r a; req = undefined e = DataSourceError $ "data source not initialized: " <> dataSourceName req Just state -> return $ wrapFetchInTrace i (length reqs) (dataSourceName (undefined :: r a)) $ wrapFetchInCatch reqs $ fetch state f (userEnv env) reqs fetches <- mapM applyFetch $ zip [n..] jobs times <- if report f >= 2 then do (refs, timedfetches) <- mapAndUnzipM wrapFetchInTimer fetches scheduleFetches timedfetches mapM (fmap Just . readIORef) refs else do scheduleFetches fetches return $ repeat Nothing let dsroundstats = HashMap.fromList [ (name, DataSourceRoundStats { dataSourceFetches = fetches , dataSourceTime = time }) | ((name, fetches), time) <- zip roundstats times] t1 <- getCurrentTime let roundtime = realToFrac (diffUTCTime t1 t0) :: Double ifReport f 1 $ modifyIORef' sref $ \(Stats rounds) -> roundstats `deepseq` Stats (RoundStats (microsecs roundtime) dsroundstats: rounds) ifTrace f 1 $ printf "Batch data fetch done (%.2fs)\n" (realToFrac roundtime :: Double) return n' -- Catch exceptions arising from the data source and stuff them into -- the appropriate requests. We don't want any exceptions propagating -- directly from the data sources, because we want the exception to be -- thrown by dataFetch instead. -- wrapFetchInCatch :: [BlockedFetch req] -> PerformFetch -> PerformFetch wrapFetchInCatch reqs fetch = case fetch of SyncFetch io -> SyncFetch (io `Exception.catch` handler) AsyncFetch fio -> AsyncFetch (\io -> fio io `Exception.catch` handler) where handler :: SomeException -> IO () handler e = do rethrowAsyncExceptions e mapM_ (forceError e) reqs -- Set the exception even if the request already had a result. -- Otherwise we could be discarding an exception. forceError e (BlockedFetch _ rvar) = do void $ tryTakeResult rvar putResult rvar (except e) wrapFetchInTimer :: PerformFetch -> IO (IORef Microseconds, PerformFetch) wrapFetchInTimer f = do r <- newIORef 0 case f of SyncFetch io -> return (r, SyncFetch (time io >>= writeIORef r)) AsyncFetch f -> do inner_r <- newIORef 0 return (r, AsyncFetch $ \inner -> do total <- time (f (time inner >>= writeIORef inner_r)) inner_t <- readIORef inner_r writeIORef r (total - inner_t)) wrapFetchInTrace :: Int -> Int -> Text.Text -> PerformFetch -> PerformFetch #ifdef EVENTLOG wrapFetchInTrace i n dsName f = case f of SyncFetch io -> SyncFetch (wrapF "Sync" io) AsyncFetch fio -> AsyncFetch (wrapF "Async" . fio . unwrapF "Async") where d = Text.unpack dsName wrapF :: String -> IO a -> IO a wrapF ty = bracket_ (traceEventIO $ printf "START %d %s (%d %s)" i d n ty) (traceEventIO $ printf "STOP %d %s (%d %s)" i d n ty) unwrapF :: String -> IO a -> IO a unwrapF ty = bracket_ (traceEventIO $ printf "STOP %d %s (%d %s)" i d n ty) (traceEventIO $ printf "START %d %s (%d %s)" i d n ty) #else wrapFetchInTrace _ _ _ f = f #endif time :: IO () -> IO Microseconds time io = do t0 <- getCurrentTime io t1 <- getCurrentTime return . microsecs . realToFrac $ t1 `diffUTCTime` t0 microsecs :: Double -> Microseconds microsecs t = round (t * 10^(6::Int)) -- | Start all the async fetches first, then perform the sync fetches before -- getting the results of the async fetches. scheduleFetches :: [PerformFetch] -> IO() scheduleFetches fetches = async_fetches sync_fetches where async_fetches :: IO () -> IO () async_fetches = compose [f | AsyncFetch f <- fetches] sync_fetches :: IO () sync_fetches = sequence_ [io | SyncFetch io <- fetches] -- ----------------------------------------------------------------------------- -- Memoization -- | A variable in the cache representing the state of a memoized computation newtype MemoVar u a = MemoVar (IORef (MemoStatus u a)) -- | The state of a memoized computation data MemoStatus u a = MemoInProgress (RoundId u) (GenHaxl u a) -- ^ Under evaluation in the given round, here is the latest -- continuation. The continuation might be a little out of -- date, but that's fine, the worst that can happen is we do a -- little extra work. | MemoDone (Either SomeException a) -- fully evaluated, here is the result. type RoundId u = IORef (RequestStore u) {- Note [RoundId] A token representing the round. This needs to be unique per round, and it needs to support Eq. Fortunately the IORef RequestStore is exactly what we need: IORef supports Eq, and we make a new one for each round. There's a danger that storing this in the DataCache could cause a space leak, so we stub out the contents after each round (see runHaxl). -} -- | 'cachedComputation' memoizes a Haxl computation. The key is a -- request. -- -- /Note:/ These cached computations will /not/ be included in the output -- of 'dumpCacheAsHaskell'. -- cachedComputation :: forall req u a. (Request req a) => req a -> GenHaxl u a -> GenHaxl u a cachedComputation req haxl = GenHaxl $ \env ref -> do cache <- readIORef (memoRef env) case DataCache.lookup req cache of Nothing -> do memovar <- newIORef (MemoInProgress ref haxl) writeIORef (memoRef env) $! DataCache.insert req (MemoVar memovar) cache run memovar haxl env ref Just (MemoVar memovar) -> do status <- readIORef memovar case status of MemoDone r -> done r MemoInProgress round cont | round == ref -> return (Blocked (retryMemo req)) | otherwise -> run memovar cont env ref -- was blocked in a previous round; run the saved continuation to -- make more progress. where -- If we got blocked on this memo in the current round, this is the -- continuation: just try to evaluate the memo again. We know it is -- already in the cache (because we just checked), so the computation -- will never be used. retryMemo req = cachedComputation req (throw (CriticalError "retryMemo")) -- Run the memoized computation and store the result (complete or -- partial) back in the MemoVar afterwards. -- -- We don't attempt to catch IO monad exceptions here. That may seem -- dangerous, because if an IO exception is raised we'll leave the -- MemoInProgress in the MemoVar. But we always want to just -- propagate an IO monad exception (it should kill the whole runHaxl, -- unless there's a unsafeToHaxlException), so we should never be -- looking at the MemoVar again anyway. Furthermore, storing the -- exception in the MemoVar is wrong, because that will turn it into -- a Haxl exception (see rethrowAsyncExceptions). run memovar cont env ref = do e <- unHaxl cont env ref case e of Done a -> complete memovar (Right a) Throw e -> complete memovar (Left e) Blocked cont -> do writeIORef memovar (MemoInProgress ref cont) return (Blocked (retryMemo req)) -- We're finished: store the final result complete memovar r = do writeIORef memovar (MemoDone r) done r -- | Lifts an 'Either' into either 'Throw' or 'Done'. done :: Either SomeException a -> IO (Result u a) done = return . either Throw Done -- ----------------------------------------------------------------------------- -- | Dump the contents of the cache as Haskell code that, when -- compiled and run, will recreate the same cache contents. For -- example, the generated code looks something like this: -- -- > loadCache :: GenHaxl u () -- > loadCache = do -- > cacheRequest (ListWombats 3) (Right ([1,2,3])) -- > cacheRequest (CountAardvarks "abcabc") (Right (2)) -- dumpCacheAsHaskell :: GenHaxl u String dumpCacheAsHaskell = do ref <- env cacheRef -- NB. cacheRef, not memoRef. We ignore memoized -- results when dumping the cache. entries <- unsafeLiftIO $ readIORef ref >>= showCache let mk_cr (req, res) = text "cacheRequest" <+> parens (text req) <+> parens (result res) result (Left e) = text "except" <+> parens (text (show e)) result (Right s) = text "Right" <+> parens (text s) return $ show $ text "loadCache :: GenHaxl u ()" $$ text "loadCache = do" $$ nest 2 (vcat (map mk_cr (concatMap snd entries))) $$ text "" -- final newline
iblumenfeld/Haxl
Haxl/Core/Monad.hs
bsd-3-clause
26,121
0
24
6,164
5,906
2,996
2,910
378
5
{-# LANGUAGE TemplateHaskell #-} -- Trac #2632 module MkData where import Language.Haskell.TH op :: Num v => v -> v -> v op a b = a + b decl1 = [d| func = 0 `op` 3 |] decl2 = [d| op x y = x func = 0 `op` 3 |]
forked-upstream-packages-for-ghcjs/ghc
testsuite/tests/th/T2632.hs
bsd-3-clause
228
0
7
71
64
40
24
8
1
import System.IO hiding (writeFile) import System.Environment import Prelude hiding (writeFile, concat) import Data.ByteString.Lazy (writeFile, concat) import Data.Binary (encode) import Control.Monad (when) import System.Exit (exitFailure) import System.FilePath ((</>)) import Language.RuScript.Serialize import Language.RuScript.Codegen import Language.RuScript.Parser import Language.RuScript.StaticCheck import Language.RuScript.Option import Language.RuScript.Optimize import Language.RuScript.Import main :: IO () main = do hSetBuffering stdout NoBuffering args <- getArgs if (length args > 1) then do let opt = parseOpt $ drop 2 args let includeDir = _includeOpt opt txt <- readFile (includeDir </> head args) let target = includeDir </> (args !! 1) case parseProgram txt of Left err -> exitError $ "Error in parsing: " ++ show err Right program -> do program' <- resolveDeps includeDir program when (_debugOpt opt) $ print program' case checkProgram program' of Left err -> exitError $ "Error in checking: " ++ err Right _ -> do let bytecode = if (_optimizeOpt opt) then optimize $ runCodegen program' else runCodegen program' when (_debugOpt opt) $ printCode bytecode writeFile target (concat $ map (encode . serialize) bytecode) else putStrLn "usage: rusc <source> <target>" where exitError s = hPutStrLn stderr s >> exitFailure
izgzhen/RuScript
rusc/src/Main.hs
mit
1,680
0
25
523
455
231
224
39
5
-- To turn a file list like: 01.first 02.second 03.third into, say: 25.first 26.second 27.third import System.Environment import Data.List import Text.Printf import System.Posix.Files import System.Directory prependNewTag :: FilePath -> FilePath -> FilePath prependNewTag orig newNum = newNum ++ ( dropWhile (/='.') orig ) myIntToTag :: Int -> FilePath myIntToTag i = printf "%.4d" i renameFiles :: [FilePath] -> Int -> [IO ()] renameFiles [] startNum = [] renameFiles (x:xs) startNum = ( rename x ( prependNewTag x $ myIntToTag startNum ) ) : renameFiles xs (startNum + 1) doFiles :: Int -> IO () doFiles startNum = do files <- fmap sort $ getDirectoryContents "." sequence_ $ renameFiles ( dropWhile ( (=='.') . head ) files ) startNum main = do args <- getArgs let num = (read::String -> Int) $ head args doFiles num
ChristopherMorton/Renumber.hs
renumber.hs
mit
848
0
13
163
295
151
144
20
1
import XMonad import System.Taffybar.Support.PagerHints (pagerHints) import XMonad.Hooks.EwmhDesktops (ewmh) import XMonad.Hooks.ManageDocks import XMonad.Util.EZConfig myKeys = [ ("<XF86AudioMute>" , spawn "amixer -c 0 set Master toggle") , ("<XF86AudioLowerVolume>", spawn "amixer -c 0 set Master 5%-") , ("<XF86AudioRaiseVolume>", spawn "amixer -c 0 set Master 5%+") , ("<XF86AudioMicMute>" , spawn "amixer -c 0 set Capture toggle") ] main = xmonad $ docks $ ewmh $ pagerHints $ def { borderWidth = 4 , modMask = mod4Mask , terminal = "termite" , layoutHook = avoidStruts $ layoutHook defaultConfig , manageHook = manageHook defaultConfig <+> manageDocks } `additionalKeysP` myKeys
rossabaker/dotfiles
xmonad/xmonad.hs
mit
775
0
10
183
165
98
67
20
1
module Data.LTE ( ) where
mckeankylej/fast-nats
src/Data/LTE.hs
mit
30
0
3
9
9
6
3
2
0
module Lesson2.BFS.Enumerator where import Control.Monad.Trans (MonadIO, liftIO) import Data.Ord (Ord(..), comparing) import qualified Data.Set as Set -------------------- import Data.Enumerator ( Stream(..) , Step(..) , Iteratee(..) , Enumerator , (>>==) , returnI , continue , yield ) import Data.Hashable (Hashable(..)) -------------------- import Navigation.Enumerator import Lesson2.Types import Lesson2.BFS.Types ------------------------------------------------------------------------------- enumBFS :: (MonadIO m, Hashable a) => Node a -> BFSGraph a -> Enumerator (NavEvent (Node a)) m b enumBFS source0 g = enumNavigation (\_ _ -> return 1) (return . (`getNodeNeighbours` g)) source0
roman/ai-class-haskell
src/Lesson2/BFS/Enumerator.hs
mit
788
0
12
176
220
134
86
25
1
{-# LANGUAGE NoImplicitPrelude, FlexibleInstances, OverloadedStrings #-} module IHaskell.Eval.Hoogle ( search, document, render, OutputFormat(..), HoogleResult, ) where import IHaskellPrelude import qualified Data.Text as T import qualified Data.Text.Lazy as LT import qualified Data.ByteString as BS import qualified Data.ByteString.Lazy as LBS import qualified Data.ByteString.Char8 as CBS import Network.HTTP.Client import Network.HTTP.Client.TLS import Data.Aeson import qualified Data.List as List import Data.Char (isAscii, isAlphaNum) import IHaskell.IPython import StringUtils (split, strip, replace) -- | Types of formats to render output to. data OutputFormat = Plain -- ^ Render to plain text. | HTML -- ^ Render to HTML. data HoogleResponse = HoogleResponse { location :: String, self :: String, docs :: String } deriving (Eq, Show) data HoogleResult = SearchResult HoogleResponse | DocResult HoogleResponse | NoResult String deriving Show data HoogleResponseList = HoogleResponseList [HoogleResponse] instance FromJSON HoogleResponseList where parseJSON (Object obj) = do results <- obj .: "results" HoogleResponseList <$> mapM parseJSON results parseJSON _ = fail "Expected object with 'results' field." instance FromJSON HoogleResponse where parseJSON (Object obj) = HoogleResponse <$> obj .: "location" <*> obj .: "self" <*> obj .: "docs" parseJSON _ = fail "Expected object with fields: location, self, docs" -- | Query Hoogle for the given string. This searches Hoogle using the internet. It returns either -- an error message or the successful JSON result. query :: String -> IO (Either String String) query str = do request <- parseUrlThrow $ queryUrl $ urlEncode str mgr <- newManager tlsManagerSettings catch (Right . CBS.unpack . LBS.toStrict . responseBody <$> httpLbs request mgr) (\e -> return $ Left $ show (e :: SomeException)) where queryUrl :: String -> String queryUrl = printf "https://www.haskell.org/hoogle/?hoogle=%s&mode=json" -- | Copied from the HTTP package. urlEncode :: String -> String urlEncode [] = [] urlEncode (ch:t) | (isAscii ch && isAlphaNum ch) || ch `elem` ("-_.~" :: String) = ch : urlEncode t | not (isAscii ch) = foldr escape (urlEncode t) (eightBs [] (fromEnum ch)) | otherwise = escape (fromEnum ch) (urlEncode t) where escape :: Int -> String -> String escape b rs = '%' : showH (b `div` 16) (showH (b `mod` 16) rs) showH :: Int -> String -> String showH x xs | x <= 9 = toEnum (o_0 + x) : xs | otherwise = toEnum (o_A + (x - 10)) : xs where o_0 = fromEnum '0' o_A = fromEnum 'A' eightBs :: [Int] -> Int -> [Int] eightBs acc x | x <= 255 = x : acc | otherwise = eightBs ((x `mod` 256) : acc) (x `div` 256) -- | Search for a query on Hoogle. Return all search results. search :: String -> IO [HoogleResult] search string = do response <- query string return $ case response of Left err -> [NoResult err] Right json -> case eitherDecode $ LBS.fromStrict $ CBS.pack json of Left err -> [NoResult err] Right results -> case map SearchResult $ (\(HoogleResponseList l) -> l) results of [] -> [NoResult "no matching identifiers found."] res -> res -- | Look up an identifier on Hoogle. Return documentation for that identifier. If there are many -- identifiers, include documentation for all of them. document :: String -> IO [HoogleResult] document string = do matchingResults <- filter matches <$> search string let results = map toDocResult matchingResults return $ case results of [] -> [NoResult "no matching identifiers found."] res -> res where matches (SearchResult resp) = case split " " $ self resp of name:_ -> strip string == strip name _ -> False matches _ = False toDocResult (SearchResult resp) = DocResult resp -- | Render a Hoogle search result into an output format. render :: OutputFormat -> HoogleResult -> String render Plain = renderPlain render HTML = renderHtml -- | Render a Hoogle result to plain text. renderPlain :: HoogleResult -> String renderPlain (NoResult res) = "No response available: " ++ res renderPlain (SearchResult resp) = printf "%s\nURL: %s\n%s" (self resp) (location resp) (docs resp) renderPlain (DocResult resp) = printf "%s\nURL: %s\n%s" (self resp) (location resp) (docs resp) -- | Render a Hoogle result to HTML. renderHtml :: HoogleResult -> String renderHtml (NoResult resp) = printf "<span class='err-msg'>No result: %s</span>" resp renderHtml (DocResult resp) = renderSelf (self resp) (location resp) ++ renderDocs (docs resp) renderHtml (SearchResult resp) = renderSelf (self resp) (location resp) ++ renderDocs (docs resp) renderSelf :: String -> String -> String renderSelf string loc | "package" `isPrefixOf` string = pkg ++ " " ++ span "hoogle-package" (link loc $ extractPackage string) | "module" `isPrefixOf` string = let package = extractPackageName loc in mod ++ " " ++ span "hoogle-module" (link loc $ extractModule string) ++ packageSub package | "class" `isPrefixOf` string = let package = extractPackageName loc in cls ++ " " ++ span "hoogle-class" (link loc $ extractClass string) ++ packageSub package | "data" `isPrefixOf` string = let package = extractPackageName loc in dat ++ " " ++ span "hoogle-class" (link loc $ extractData string) ++ packageSub package | otherwise = let [name, args] = split "::" string package = extractPackageName loc modname = extractModuleName loc in span "hoogle-name" (unicodeReplace $ link loc (strip name) ++ " :: " ++ strip args) ++ packageAndModuleSub package modname where extractPackage = strip . replace "package" "" extractModule = strip . replace "module" "" extractClass = strip . replace "class" "" extractData = strip . replace "data" "" pkg = span "hoogle-head" "package" mod = span "hoogle-head" "module" cls = span "hoogle-head" "class" dat = span "hoogle-head" "data" unicodeReplace :: String -> String unicodeReplace = replace "forall" "&#x2200;" . replace "=>" "&#x21D2;" . replace "->" "&#x2192;" . replace "::" "&#x2237;" packageSub Nothing = "" packageSub (Just package) = span "hoogle-sub" $ "(" ++ pkg ++ " " ++ span "hoogle-package" package ++ ")" packageAndModuleSub Nothing _ = "" packageAndModuleSub (Just package) Nothing = packageSub (Just package) packageAndModuleSub (Just package) (Just modname) = span "hoogle-sub" $ "(" ++ pkg ++ " " ++ span "hoogle-package" package ++ ", " ++ mod ++ " " ++ span "hoogle-module" modname ++ ")" renderDocs :: String -> String renderDocs doc = let groups = List.groupBy bothAreCode $ lines doc nonull = filter (not . null . strip) bothAreCode s1 s2 = isPrefixOf ">" (strip s1) && isPrefixOf ">" (strip s2) isCode (s:_) = isPrefixOf ">" $ strip s makeBlock lines = if isCode lines then div' "hoogle-code" $ unlines $ nonull lines else div' "hoogle-text" $ unlines $ nonull lines in div' "hoogle-doc" $ unlines $ map makeBlock groups extractPackageName :: String -> Maybe String extractPackageName link = do let pieces = split "/" link archiveLoc <- List.elemIndex "archive" pieces latestLoc <- List.elemIndex "latest" pieces guard $ latestLoc - archiveLoc == 2 return $ pieces List.!! (latestLoc - 1) extractModuleName :: String -> Maybe String extractModuleName link = do let pieces = split "/" link guard $ not $ null pieces let html = fromJust $ lastMay pieces mod = replace "-" "." $ takeWhile (/= '.') html return mod div' :: String -> String -> String div' = printf "<div class='%s'>%s</div>" span :: String -> String -> String span = printf "<span class='%s'>%s</span>" link :: String -> String -> String link = printf "<a target='_blank' href='%s'>%s</a>"
sumitsahrawat/IHaskell
src/IHaskell/Eval/Hoogle.hs
mit
8,550
0
20
2,263
2,518
1,267
1,251
199
4
module Error where import LispData import Control.Monad.Except import Text.ParserCombinators.Parsec (ParseError) data LispError = NumArgs Integer [LispVal] | TypeMismatch String LispVal | Parser ParseError | BadSpecialForm String LispVal | NotFunction String String | UnboundVar String String | Default String showError :: LispError -> String showError (UnboundVar message varname) = message ++ ": " ++ varname showError (BadSpecialForm message form) = message ++ ": " ++ show form showError (NotFunction message func) = message ++ ": " ++ show func showError (NumArgs expected found) = "Expected " ++ show expected ++ " args; found values " ++ unwordsList found showError (TypeMismatch expected found) = "Invalid type: expected " ++ expected ++ ", found " ++ show found showError (Parser parseErr) = "Parse error at " ++ show parseErr instance Show LispError where show = showError type ThrowsError = Either LispError trapError :: (Either LispError String) -> Either LispError String trapError action = catchError action (return . show) extractValue :: ThrowsError a -> a extractValue (Right val) = val
evinstk/scheme-by-haskell
Error.hs
mit
1,222
0
8
284
339
176
163
24
1
-- Game State module State(State(..), initialGameState, resetGameState) where import Playfield import Piece import System.Random data State = State { well :: Well , time :: Float , deltaTime :: Float , secondsToNextMove :: Float , piece :: Piece , piecePos :: (Int, Int) , randomSeed :: StdGen , score :: Int , accelerate :: Bool } deriving (Show) initialGameState :: State initialGameState = State { well = emptyWell , time = 0 , deltaTime = 0 , secondsToNextMove = 0 , piece = tetrominoO , piecePos = (0, 0) , randomSeed = mkStdGen 0 -- found better way! , score = 0 , accelerate = False } -- Resets a game state, maintaining the random seed resetGameState :: State -> State resetGameState s = initialGameState {randomSeed = (randomSeed s)}
mgeorgoulopoulos/TetrisHaskellWeekend
State.hs
mit
866
0
9
252
215
137
78
28
1
module Main where import Git main :: IO () main = do putStrLn <- currentBranch putStrLn <- remoteUrl return ()
roberto/octo-reviews
src/Main.hs
mit
119
0
8
28
44
22
22
7
1
{-# LANGUAGE OverloadedStrings #-} module JoScript.Data.Error where import Protolude hiding (Location) import Data.Aeson ((.=)) import qualified Data.Aeson.Types as A import JoScript.Data.Position (Position) import JoScript.Util.Json (withObject) import JoScript.Data.Lexer (LpRepr(..), LpReprKind(..)) import qualified JoScript.Data.Block as Bp import qualified JoScript.Data.Lexer as Lp data Error = Error Repr Location deriving (Eq, Show) data Location = Known Position deriving (Eq, Show) data Label = Label { position :: Position, description :: Text } deriving (Show, Eq) data Repr = IndentError IndentErrorT | LexerError LexerErrorT | ParseError ParseErrorT [Label] deriving (Eq, Show) data IndentErrorT = ShallowDedent deriving (Eq, Show) data LexerErrorT = LUnexpectedEnd | LUnexpectedToken Bp.BpRepr | LUnexpectedCharacter Char | LUnknownTokenStart Char | LInvalidIntSuffix Char | LDuplicateDecimial deriving (Eq, Show) data ParseErrorT = PUnexpectedEnd | PImpossible | PIncompleteAlt | PExpectedToken LpReprKind LpRepr | PUnexpectedToken LpRepr deriving (Eq, Show) -------------------------------------------------------------- -- exports -- -------------------------------------------------------------- known :: Repr -> Position -> Error known k p = Error k (Known p) reprDescription :: Repr -> Text reprDescription (IndentError _) = "text:block" reprDescription (LexerError _) = "text:lexer" reprDescription (ParseError _ _) = "text:parse" lexerErrMsg :: LexerErrorT -> Text lexerErrMsg LUnexpectedEnd = "unexpected lexer ended" lexerErrMsg (LUnexpectedToken _) = "unexpected block token" lexerErrMsg (LUnexpectedCharacter _) = "unexpected character" lexerErrMsg (LUnknownTokenStart _) = "unexpected character" lexerErrMsg (LInvalidIntSuffix _) = "integer was suffixed with invalid character" lexerErrMsg LDuplicateDecimial = "duplicated decimal place in float" parseErrMsg :: ParseErrorT -> Text parseErrMsg PUnexpectedEnd = "unexpected parse end during parse" parseErrMsg PIncompleteAlt = "implementation error" parseErrMsg PUnexpectedToken{} = "encounted unexpected token" parseErrMsg PExpectedToken{} = "Expected different token" parseErrMsg PImpossible = "Impossible error" {- Determines which error is most recently occuring in a file -} newestError :: Error -> Error -> Error newestError a@(Error _ (Known al)) b@(Error _ (Known bl)) | al > bl = a | bl > al = b | otherwise = a -------------------------------------------------------------- -- instances -- -------------------------------------------------------------- instance A.ToJSON Error where toJSON (Error repr loc) = A.object ["location" .= loc, "repr" .= repr] instance A.ToJSON Location where toJSON (Known p) = withObject ["type" .= knownS] (A.toJSON p) where knownS = "known" :: Text instance A.ToJSON Label where toJSON Label{..} = A.object ["description" .= description, "position" .= position] instance A.ToJSON Repr where toJSON t = withObject ["level" .= reprDescription t] (withType t) where withType (IndentError err) = A.toJSON err withType (LexerError err) = A.toJSON err withType (ParseError err lbs) = withObject ["path" .= lbs] (A.toJSON err) instance A.ToJSON IndentErrorT where toJSON ShallowDedent = A.object ["message" .= ("dedent depth is too shallow" :: Text)] instance A.ToJSON LexerErrorT where toJSON t = withObject ["message" .= lexerErrMsg t] (withType t) where withType (LUnexpectedToken token) = A.object ["token" .= token] withType (LUnknownTokenStart c) = A.object ["character" .= c] withType (LUnexpectedCharacter c) = A.object ["character" .= c] withType (LInvalidIntSuffix c) = A.object ["character" .= c] withType _ = A.object [] instance A.ToJSON ParseErrorT where toJSON t = withObject ["message" .= parseErrMsg t] (withType t) where withType (PUnexpectedToken t) = A.object ["read" .= t] withType (PExpectedToken e t) = A.object ["read" .= t, "expected-type" .= e] withType ____________________ = A.object []
AKST/jo
source/lib/JoScript/Data/Error.hs
mit
4,220
0
11
795
1,197
638
559
-1
-1
data ArrowOpts n = ArrowOpts { _arrowHead :: ArrowHT n , _arrowTail :: ArrowHT n , _arrowShaft :: Trail V2 n , _headGap :: Measure n , _tailGap :: Measure n , _headStyle :: Style V2 n , _headLength :: Measure n , _tailStyle :: Style V2 n , _tailLength :: Measure n , _shaftStyle :: Style V2 n }
jeffreyrosenbluth/NYC-meetup
meetup/ArrowOpts.hs
mit
351
0
9
116
109
60
49
12
0
import Probs1to10 import Probs11to20 import Probs21to30 import Probs31to40 main :: IO () main = do putStrLn $ show $ mlast "RandPer" putStrLn $ show $ lastbutone "RandPer" putStrLn $ show $ kth "RandPer" 0 putStrLn $ show $ mlen "RandPer" putStrLn $ show $ rev "RandPer" putStrLn $ show $ frev "RandPer" putStrLn $ show $ ispalin "RandPer" putStrLn $ show $ flatlist ((List [Elem 1, List [Elem 2, List [Elem 3, Elem 4], Elem 5]]) :: NestedList Int) putStrLn $ show $ flatten ((List [Elem 1, List [Elem 2, List [Elem 3, Elem 4], Elem 5]]) :: NestedList Int) putStrLn $ show $ remdups "aaaabccaadeeee" putStrLn $ show $ remdups' "aaaabccaadeeee" putStrLn $ show $ pack "aaaabccaadeeee" putStrLn $ show $ encode $ pack "aaaabccaadeeee" putStrLn $ show $ encodemodified "aaaabccaadeeee" putStrLn $ show $ decodemodified [Multiple 4 'a',Single 'b',Multiple 2 'c', Multiple 2 'a',Single 'd',Multiple 4 'e'] putStrLn $ show $ insertat "chandra" 'M' 3
ajjaic/Problems99
src/Main.hs
mit
1,011
0
17
224
420
197
223
22
1
module Utils where import Data.Aeson import qualified Data.ByteString.Lazy as B import Control.Applicative loadJSON :: (FromJSON a) => FilePath -> IO a loadJSON a = do x <- B.readFile a >>= return . decode case x of Nothing -> error $ "Cannot load from " ++ a Just y -> return y writeJSON :: (ToJSON a) => FilePath -> a -> IO () writeJSON f = B.writeFile f . encode
edwardwas/haskcal
src/Utils.hs
mit
375
0
11
78
149
77
72
12
2
{-Generation of pink noise roughly based on the algorithm from here: http://www.firstpr.com.au/dsp/pink-noise/phil_burk_19990905_patest_pink.c From here: http://www.firstpr.com.au/dsp/pink-noise/ -} module PinkVMBurk (genPink,initialPinkNoise) where import WhiteNoise (whiteRandom) import Data.Bits trailingZeros:: Int -> Int trailingZeros n | n == 0 = 0 | n .&. 1 /= 0 = 0 | otherwise = 1 + trailingZeros (n `shiftR` 1) numOfPinkRows::Int numOfPinkRows = 30 indexMask::Integer indexMask = (1 `shiftL` numOfPinkRows) - 1 initialPinkNoise :: PinkNoise initialPinkNoise = PinkNoise initRows 0 0 0 initRows::[Double] initRows = replicate numOfPinkRows 0.0 data PinkNoise = PinkNoise { rows::[Double] , runningSum::Double , index::Int , rand2::Double } deriving (Show,Eq) data RandomValue = RandomValue { rval1::Double , rval2::Double } deriving (Show,Eq) randomValues :: Int -> Int -> [RandomValue] randomValues rate dur = map buildVal (zip3 [0..(rate * dur)] (whiteRandom 100) (whiteRandom 200)) where buildVal ( _ ,val1',val2') = RandomValue (val1' * 0.1) (val2' * 0.1) getPink::PinkNoise -> Int-> Int -> [PinkNoise] getPink pn rate dur = scanl nextpValue pn (randomValues rate dur) -- generate list of pink floating values genPink :: PinkNoise -> Int -> Int -> [Double] genPink pn rate dur = map calcVal (getPink pn rate dur) where calcVal pn'= runningSum pn' + rand2 pn' updatePink :: PinkNoise-> Int -> Double->Double -> PinkNoise updatePink pn idx rnd = PinkNoise updateRow updateRunningSum updateIndex where updateRunningSum = runningSum pn + rnd - ( rows pn!!idx) updateRow = let sp = splitAt idx (rows pn) in case fst sp of [] -> rnd:snd sp _ -> init (fst sp) ++ [rnd] ++ snd sp updateIndex = (index pn + 1) .&. fromIntegral indexMask nextpValue::PinkNoise -> RandomValue -> PinkNoise nextpValue pn rv | nextIdx /= 0 = updatePink pn (trailingZeros nextIdx) (rval1 rv) (rval2 rv) | otherwise = pn {index = nextIdx,rand2 = rval2 rv} where nextIdx = index pn + 1 .&. fromIntegral indexMask
danplubell/color-noise
library/PinkVMBurk.hs
mit
2,590
0
16
897
737
387
350
51
2
fibs :: [Integer] fibs = 1:2:zipWith (+) fibs (tail fibs) main = print . sum . takeWhile (<4000000) . filter even $ fibs
chrisfosterelli/euler.hs
solutions/002.hs
mit
121
0
9
23
68
36
32
3
1
{-# LANGUAGE PatternSynonyms #-} -- For HasCallStack compatibility {-# LANGUAGE ImplicitParams, ConstraintKinds, KindSignatures #-} {-# OPTIONS_GHC -fno-warn-unused-imports #-} module JSDOM.Generated.SVGNumber (setValue, getValue, SVGNumber(..), gTypeSVGNumber) where import Prelude ((.), (==), (>>=), return, IO, Int, Float, Double, Bool(..), Maybe, maybe, fromIntegral, round, realToFrac, fmap, Show, Read, Eq, Ord, Maybe(..)) import qualified Prelude (error) import Data.Typeable (Typeable) import Data.Traversable (mapM) import Language.Javascript.JSaddle (JSM(..), JSVal(..), JSString, strictEqual, toJSVal, valToStr, valToNumber, valToBool, js, jss, jsf, jsg, function, asyncFunction, new, array, jsUndefined, (!), (!!)) import Data.Int (Int64) import Data.Word (Word, Word64) import JSDOM.Types import Control.Applicative ((<$>)) import Control.Monad (void) import Control.Lens.Operators ((^.)) import JSDOM.EventTargetClosures (EventName, unsafeEventName, unsafeEventNameAsync) import JSDOM.Enums -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGNumber.value Mozilla SVGNumber.value documentation> setValue :: (MonadDOM m) => SVGNumber -> Float -> m () setValue self val = liftDOM (self ^. jss "value" (toJSVal val)) -- | <https://developer.mozilla.org/en-US/docs/Web/API/SVGNumber.value Mozilla SVGNumber.value documentation> getValue :: (MonadDOM m) => SVGNumber -> m Float getValue self = liftDOM (realToFrac <$> ((self ^. js "value") >>= valToNumber))
ghcjs/jsaddle-dom
src/JSDOM/Generated/SVGNumber.hs
mit
1,486
0
12
171
414
260
154
23
1
module RotButtons where import Control.Monad import Control.Monad.Instances import Data.IORef import Graphics.UI.Gtk hiding (on,get) import qualified Graphics.UI.Gtk as G import Graphics.UI.Gtk.Glade import Numeric.LinearAlgebra ((<>),(><),(|>),(@>),(@@>),Vector,Matrix) import qualified Numeric.LinearAlgebra as V import Types import Utils refreshButtons :: IORef GameState -> IO () refreshButtons gameStateRef = do gameState <- readIORef gameStateRef let (up,down,left,right) = getButtonBoxes gameState n = getN gameState rotStep = getRotStep gameState drawingArea = getDrawingArea gameState xml = getXML gameState [upB,downB,leftB,rightB] <- mapM (mapM $ (\ k -> fmap ((,) k) $ buttonNewWithLabel $ show k)) $ replicate 4 [3..n] containerForeach up (containerRemove up) containerForeach down (containerRemove down) containerForeach left (containerRemove left) containerForeach right (containerRemove right) mapM_ (processButton gameStateRef up 1) upB mapM_ (processButton gameStateRef down 3) downB mapM_ (processButton gameStateRef left 2) leftB mapM_ (processButton gameStateRef right 0) rightB buttonClickAction xml "rotCW" (buttonAction gameStateRef 4 0) buttonClickAction xml "rotCCW" (buttonAction gameStateRef 5 0) return () processButton :: (ContainerClass b) => IORef GameState -> b -> Int -> (Int,Button) -> IO () processButton g box side (k,button) = do -- side: 0 = right, going CCW containerAdd box button widgetShow button G.on button buttonActivated $ buttonAction g side k return () buttonAction :: IORef GameState -> Int -> Int -> IO () buttonAction gameStateRef side k = do gameState <- readIORef gameStateRef let rotStep = getRotStep gameState n = getN gameState d = getDrawingArea gameState rp = case side of 0 -> rotPlane n 0 (k-1) . negate 1 -> rotPlane n 1 (k-1) . negate 2 -> rotPlane n 0 (k-1) 3 -> rotPlane n 1 (k-1) 4 -> rotPlane n 0 1 5 -> rotPlane n 0 1 . negate modifyIORef gameStateRef (\gS -> gS { getViewPoint = rp rotStep <> getViewPoint gS }) widgetQueueDraw d
benjaminjkraft/nmines
RotButtons.hs
mit
2,190
0
19
477
795
403
392
52
6