Search is not available for this dataset
repo_name
string | path
string | license
string | full_code
string | full_size
int64 | uncommented_code
string | uncommented_size
int64 | function_only_code
string | function_only_size
int64 | is_commented
bool | is_signatured
bool | n_ast_errors
int64 | ast_max_depth
int64 | n_whitespaces
int64 | n_ast_nodes
int64 | n_ast_terminals
int64 | n_ast_nonterminals
int64 | loc
int64 | cycloplexity
int64 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
erantapaa/read-word32
|
src/Lib.hs
|
bsd-3-clause
|
wanted :: BL.ByteString -> Bool
wanted bs = BL.index bs 0 == 0xC0
| 65
|
wanted :: BL.ByteString -> Bool
wanted bs = BL.index bs 0 == 0xC0
| 65
|
wanted bs = BL.index bs 0 == 0xC0
| 33
| false
| true
| 0
| 7
| 12
| 38
| 16
| 22
| null | null |
alexander-at-github/eta
|
compiler/ETA/Types/TyCon.hs
|
bsd-3-clause
|
-- | Kind constructors
mkKindTyCon :: Name -> Kind -> TyCon
mkKindTyCon name kind
= mkPrimTyCon' name kind [] VoidRep True
| 124
|
mkKindTyCon :: Name -> Kind -> TyCon
mkKindTyCon name kind
= mkPrimTyCon' name kind [] VoidRep True
| 101
|
mkKindTyCon name kind
= mkPrimTyCon' name kind [] VoidRep True
| 64
| true
| true
| 0
| 6
| 22
| 38
| 19
| 19
| null | null |
grtlr/wyas
|
src/Wyas/Parser.hs
|
bsd-3-clause
|
parseQuasiQuoted :: Parser LispVal
parseQuasiQuoted = do char '`'
x <- parseExpr
return $ List [Atom "quasiquote", x]
| 161
|
parseQuasiQuoted :: Parser LispVal
parseQuasiQuoted = do char '`'
x <- parseExpr
return $ List [Atom "quasiquote", x]
| 161
|
parseQuasiQuoted = do char '`'
x <- parseExpr
return $ List [Atom "quasiquote", x]
| 126
| false
| true
| 0
| 10
| 61
| 46
| 21
| 25
| null | null |
asilvestre/haskell-neo4j-rest-client
|
tests/IntegrationTests.hs
|
mit
|
case_uniquenessTraversalNodes :: Assertion
case_uniquenessTraversalNodes = withAuthConnection host port creds $ do
g <- setUpTraversalTest
let start = fromJust $ G.getNamedNode "Tobias" g
let desc = def {T.travDepth = Left 2, T.travRelFilter = [("loves", Any), ("hates", Any), ("admires", Any)]}
ns <- T.traverseGetNodes desc start
let expected = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected) (L.sort ns)
let desc2 = desc {T.travUniqueness = Just T.NodeGlobal}
ns2 <- T.traverseGetNodes desc2 start
let expected2 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected2) (L.sort ns2)
let desc3 = desc {T.travUniqueness = Just T.RelationshipGlobal}
ns3 <- T.traverseGetNodes desc3 start
let expected3 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected3) (L.sort ns3)
let desc4 = desc {T.travUniqueness = Just T.NodePathUnique}
ns4 <- T.traverseGetNodes desc4 start
let expected4 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected4) (L.sort ns4)
let desc5 = desc {T.travUniqueness = Just T.RelationshipPath}
ns5 <- T.traverseGetNodes desc5 start
let expected5 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected5) (L.sort ns5)
cleanUpTraversalTest g
-- | Test traversal with a javascript depth expression
| 1,643
|
case_uniquenessTraversalNodes :: Assertion
case_uniquenessTraversalNodes = withAuthConnection host port creds $ do
g <- setUpTraversalTest
let start = fromJust $ G.getNamedNode "Tobias" g
let desc = def {T.travDepth = Left 2, T.travRelFilter = [("loves", Any), ("hates", Any), ("admires", Any)]}
ns <- T.traverseGetNodes desc start
let expected = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected) (L.sort ns)
let desc2 = desc {T.travUniqueness = Just T.NodeGlobal}
ns2 <- T.traverseGetNodes desc2 start
let expected2 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected2) (L.sort ns2)
let desc3 = desc {T.travUniqueness = Just T.RelationshipGlobal}
ns3 <- T.traverseGetNodes desc3 start
let expected3 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected3) (L.sort ns3)
let desc4 = desc {T.travUniqueness = Just T.NodePathUnique}
ns4 <- T.traverseGetNodes desc4 start
let expected4 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected4) (L.sort ns4)
let desc5 = desc {T.travUniqueness = Just T.RelationshipPath}
ns5 <- T.traverseGetNodes desc5 start
let expected5 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected5) (L.sort ns5)
cleanUpTraversalTest g
-- | Test traversal with a javascript depth expression
| 1,643
|
case_uniquenessTraversalNodes = withAuthConnection host port creds $ do
g <- setUpTraversalTest
let start = fromJust $ G.getNamedNode "Tobias" g
let desc = def {T.travDepth = Left 2, T.travRelFilter = [("loves", Any), ("hates", Any), ("admires", Any)]}
ns <- T.traverseGetNodes desc start
let expected = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected) (L.sort ns)
let desc2 = desc {T.travUniqueness = Just T.NodeGlobal}
ns2 <- T.traverseGetNodes desc2 start
let expected2 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected2) (L.sort ns2)
let desc3 = desc {T.travUniqueness = Just T.RelationshipGlobal}
ns3 <- T.traverseGetNodes desc3 start
let expected3 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected3) (L.sort ns3)
let desc4 = desc {T.travUniqueness = Just T.NodePathUnique}
ns4 <- T.traverseGetNodes desc4 start
let expected4 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected4) (L.sort ns4)
let desc5 = desc {T.travUniqueness = Just T.RelationshipPath}
ns5 <- T.traverseGetNodes desc5 start
let expected5 = map (\n -> fromJust $ G.getNamedNode n g) ["Tobias", "Sara", "Gumersindo", "Sara", "Gumersindo"]
neo4jEqual (L.sort expected5) (L.sort ns5)
cleanUpTraversalTest g
-- | Test traversal with a javascript depth expression
| 1,600
| false
| true
| 0
| 16
| 300
| 630
| 317
| 313
| null | null |
fros1y/umbral
|
src/Coord.hs
|
bsd-3-clause
|
insetBounds :: Integer -> Bounds -> Bounds
insetBounds i (Bounds l u) = (Bounds l' u')
where
offset = (Coord i i)
l' = l + offset
u' = u - offset
| 159
|
insetBounds :: Integer -> Bounds -> Bounds
insetBounds i (Bounds l u) = (Bounds l' u')
where
offset = (Coord i i)
l' = l + offset
u' = u - offset
| 159
|
insetBounds i (Bounds l u) = (Bounds l' u')
where
offset = (Coord i i)
l' = l + offset
u' = u - offset
| 116
| false
| true
| 2
| 7
| 45
| 80
| 38
| 42
| null | null |
coolhacks/scripts-hacks
|
examples/shellcheck-master/ShellCheck/Parser.hs
|
mit
|
parseProblem level code msg = do
pos <- getPosition
parseProblemAt pos level code msg
| 93
|
parseProblem level code msg = do
pos <- getPosition
parseProblemAt pos level code msg
| 93
|
parseProblem level code msg = do
pos <- getPosition
parseProblemAt pos level code msg
| 93
| false
| false
| 0
| 7
| 21
| 32
| 14
| 18
| null | null |
jgoerzen/dtmconv
|
HaXml-1.12/src/Text/XML/HaXml/Combinators.hs
|
gpl-2.0
|
-- FILTER COMBINATORS
-- | Sequential (/Irish/,/backwards/) composition
o :: CFilter -> CFilter -> CFilter
f `o` g = concatMap f . g
| 134
|
o :: CFilter -> CFilter -> CFilter
f `o` g = concatMap f . g
| 60
|
f `o` g = concatMap f . g
| 25
| true
| true
| 0
| 6
| 24
| 43
| 21
| 22
| null | null |
leshchevds/ganeti
|
src/Ganeti/Constants.hs
|
bsd-2-clause
|
exportModeRemote :: String
exportModeRemote = Types.exportModeToRaw ExportModeRemote
| 84
|
exportModeRemote :: String
exportModeRemote = Types.exportModeToRaw ExportModeRemote
| 84
|
exportModeRemote = Types.exportModeToRaw ExportModeRemote
| 57
| false
| true
| 0
| 6
| 6
| 16
| 8
| 8
| null | null |
nushio3/ghc
|
compiler/deSugar/TmOracle.hs
|
bsd-3-clause
|
exprDeepLookup env (PmExprEq e1 e2) = PmExprEq (exprDeepLookup env e1)
(exprDeepLookup env e2)
| 141
|
exprDeepLookup env (PmExprEq e1 e2) = PmExprEq (exprDeepLookup env e1)
(exprDeepLookup env e2)
| 141
|
exprDeepLookup env (PmExprEq e1 e2) = PmExprEq (exprDeepLookup env e1)
(exprDeepLookup env e2)
| 141
| false
| false
| 0
| 7
| 59
| 42
| 19
| 23
| null | null |
lpsmith/aeson
|
benchmarks/CompareWithJSON.hs
|
bsd-3-clause
|
decodeJ :: String -> J.JSValue
decodeJ s =
case J.decodeStrict s of
J.Ok v -> v
J.Error _ -> error "fail to parse via JSON"
| 133
|
decodeJ :: String -> J.JSValue
decodeJ s =
case J.decodeStrict s of
J.Ok v -> v
J.Error _ -> error "fail to parse via JSON"
| 133
|
decodeJ s =
case J.decodeStrict s of
J.Ok v -> v
J.Error _ -> error "fail to parse via JSON"
| 102
| false
| true
| 0
| 9
| 34
| 54
| 25
| 29
| null | null |
wavewave/xournal-types
|
src/Data/Xournal/Predefined.hs
|
bsd-2-clause
|
predefined_highlighter_opacity :: Double
predefined_highlighter_opacity = 0.5
| 78
|
predefined_highlighter_opacity :: Double
predefined_highlighter_opacity = 0.5
| 77
|
predefined_highlighter_opacity = 0.5
| 36
| false
| true
| 0
| 4
| 6
| 11
| 6
| 5
| null | null |
hecrj/haskell-format
|
src/Language/Haskell/Format/Declaration.hs
|
bsd-3-clause
|
assertion :: Asst CommentedSrc -> Format
assertion asst =
case asst of
ClassA _ qname types ->
Format.intercalate " " (Atom.qname qname : map type' types)
ParenA _ parenAsst ->
"(" <> assertion parenAsst <> ")"
_ ->
error (show asst)
| 299
|
assertion :: Asst CommentedSrc -> Format
assertion asst =
case asst of
ClassA _ qname types ->
Format.intercalate " " (Atom.qname qname : map type' types)
ParenA _ parenAsst ->
"(" <> assertion parenAsst <> ")"
_ ->
error (show asst)
| 299
|
assertion asst =
case asst of
ClassA _ qname types ->
Format.intercalate " " (Atom.qname qname : map type' types)
ParenA _ parenAsst ->
"(" <> assertion parenAsst <> ")"
_ ->
error (show asst)
| 258
| false
| true
| 6
| 12
| 106
| 97
| 46
| 51
| null | null |
randen/cabal
|
Cabal/Distribution/Simple/Build.hs
|
bsd-3-clause
|
-- | Translate a lib-style 'TestSuite' component into a lib + exe for building
testSuiteLibV09AsLibAndExe :: PackageDescription
-> TestSuite
-> ComponentLocalBuildInfo
-> LocalBuildInfo
-> FilePath
-> FilePath
-> (PackageDescription,
Library, ComponentLocalBuildInfo,
LocalBuildInfo,
IPI.InstalledPackageInfo,
Executable, ComponentLocalBuildInfo)
testSuiteLibV09AsLibAndExe pkg_descr
test@TestSuite { testInterface = TestSuiteLibV09 _ m }
clbi lbi distPref pwd =
(pkg, lib, libClbi, lbi, ipi, exe, exeClbi)
where
bi = testBuildInfo test
lib = Library {
exposedModules = [ m ],
reexportedModules = [],
requiredSignatures = [],
exposedSignatures = [],
libExposed = True,
libBuildInfo = bi
}
libClbi = LibComponentLocalBuildInfo
{ componentPackageDeps = componentPackageDeps clbi
, componentPackageRenaming = componentPackageRenaming clbi
, componentId = ComponentId $ display (packageId pkg)
, componentCompatPackageKey = ComponentId $ display (packageId pkg)
, componentExposedModules = [IPI.ExposedModule m Nothing Nothing]
}
pkg = pkg_descr {
package = (package pkg_descr) {
pkgName = PackageName (testName test)
}
, buildDepends = targetBuildDepends $ testBuildInfo test
, executables = []
, testSuites = []
, library = Just lib
}
ipi = inplaceInstalledPackageInfo pwd distPref pkg (IPI.AbiHash "") lib lbi libClbi
testDir = buildDir lbi </> stubName test
</> stubName test ++ "-tmp"
testLibDep = thisPackageVersion $ package pkg
exe = Executable {
exeName = stubName test,
modulePath = stubFilePath test,
buildInfo = (testBuildInfo test) {
hsSourceDirs = [ testDir ],
targetBuildDepends = testLibDep
: (targetBuildDepends $ testBuildInfo test),
targetBuildRenaming = Map.empty
}
}
-- | The stub executable needs a new 'ComponentLocalBuildInfo'
-- that exposes the relevant test suite library.
exeClbi = ExeComponentLocalBuildInfo {
componentPackageDeps =
(IPI.installedComponentId ipi, packageId ipi)
: (filter (\(_, x) -> let PackageName name = pkgName x
in name == "Cabal" || name == "base")
(componentPackageDeps clbi)),
componentPackageRenaming =
Map.insert (packageName ipi) defaultRenaming
(componentPackageRenaming clbi)
}
| 3,233
|
testSuiteLibV09AsLibAndExe :: PackageDescription
-> TestSuite
-> ComponentLocalBuildInfo
-> LocalBuildInfo
-> FilePath
-> FilePath
-> (PackageDescription,
Library, ComponentLocalBuildInfo,
LocalBuildInfo,
IPI.InstalledPackageInfo,
Executable, ComponentLocalBuildInfo)
testSuiteLibV09AsLibAndExe pkg_descr
test@TestSuite { testInterface = TestSuiteLibV09 _ m }
clbi lbi distPref pwd =
(pkg, lib, libClbi, lbi, ipi, exe, exeClbi)
where
bi = testBuildInfo test
lib = Library {
exposedModules = [ m ],
reexportedModules = [],
requiredSignatures = [],
exposedSignatures = [],
libExposed = True,
libBuildInfo = bi
}
libClbi = LibComponentLocalBuildInfo
{ componentPackageDeps = componentPackageDeps clbi
, componentPackageRenaming = componentPackageRenaming clbi
, componentId = ComponentId $ display (packageId pkg)
, componentCompatPackageKey = ComponentId $ display (packageId pkg)
, componentExposedModules = [IPI.ExposedModule m Nothing Nothing]
}
pkg = pkg_descr {
package = (package pkg_descr) {
pkgName = PackageName (testName test)
}
, buildDepends = targetBuildDepends $ testBuildInfo test
, executables = []
, testSuites = []
, library = Just lib
}
ipi = inplaceInstalledPackageInfo pwd distPref pkg (IPI.AbiHash "") lib lbi libClbi
testDir = buildDir lbi </> stubName test
</> stubName test ++ "-tmp"
testLibDep = thisPackageVersion $ package pkg
exe = Executable {
exeName = stubName test,
modulePath = stubFilePath test,
buildInfo = (testBuildInfo test) {
hsSourceDirs = [ testDir ],
targetBuildDepends = testLibDep
: (targetBuildDepends $ testBuildInfo test),
targetBuildRenaming = Map.empty
}
}
-- | The stub executable needs a new 'ComponentLocalBuildInfo'
-- that exposes the relevant test suite library.
exeClbi = ExeComponentLocalBuildInfo {
componentPackageDeps =
(IPI.installedComponentId ipi, packageId ipi)
: (filter (\(_, x) -> let PackageName name = pkgName x
in name == "Cabal" || name == "base")
(componentPackageDeps clbi)),
componentPackageRenaming =
Map.insert (packageName ipi) defaultRenaming
(componentPackageRenaming clbi)
}
| 3,154
|
testSuiteLibV09AsLibAndExe pkg_descr
test@TestSuite { testInterface = TestSuiteLibV09 _ m }
clbi lbi distPref pwd =
(pkg, lib, libClbi, lbi, ipi, exe, exeClbi)
where
bi = testBuildInfo test
lib = Library {
exposedModules = [ m ],
reexportedModules = [],
requiredSignatures = [],
exposedSignatures = [],
libExposed = True,
libBuildInfo = bi
}
libClbi = LibComponentLocalBuildInfo
{ componentPackageDeps = componentPackageDeps clbi
, componentPackageRenaming = componentPackageRenaming clbi
, componentId = ComponentId $ display (packageId pkg)
, componentCompatPackageKey = ComponentId $ display (packageId pkg)
, componentExposedModules = [IPI.ExposedModule m Nothing Nothing]
}
pkg = pkg_descr {
package = (package pkg_descr) {
pkgName = PackageName (testName test)
}
, buildDepends = targetBuildDepends $ testBuildInfo test
, executables = []
, testSuites = []
, library = Just lib
}
ipi = inplaceInstalledPackageInfo pwd distPref pkg (IPI.AbiHash "") lib lbi libClbi
testDir = buildDir lbi </> stubName test
</> stubName test ++ "-tmp"
testLibDep = thisPackageVersion $ package pkg
exe = Executable {
exeName = stubName test,
modulePath = stubFilePath test,
buildInfo = (testBuildInfo test) {
hsSourceDirs = [ testDir ],
targetBuildDepends = testLibDep
: (targetBuildDepends $ testBuildInfo test),
targetBuildRenaming = Map.empty
}
}
-- | The stub executable needs a new 'ComponentLocalBuildInfo'
-- that exposes the relevant test suite library.
exeClbi = ExeComponentLocalBuildInfo {
componentPackageDeps =
(IPI.installedComponentId ipi, packageId ipi)
: (filter (\(_, x) -> let PackageName name = pkgName x
in name == "Cabal" || name == "base")
(componentPackageDeps clbi)),
componentPackageRenaming =
Map.insert (packageName ipi) defaultRenaming
(componentPackageRenaming clbi)
}
| 2,600
| true
| true
| 0
| 16
| 1,381
| 588
| 325
| 263
| null | null |
plow-technologies/distributed-simple-cell
|
src/SimpleStore/Cell/Distributed/Migration.hs
|
mit
|
-- | Insert a new store with a value, or overwrite the stored value if the value's key is already present in the cell
getOrInsertStore :: (SimpleCellConstraint cell k src dst tm st) => st -> DistributedHandlerM urllist st ()
getOrInsertStore state = do
localCell <- asks localCell
(liftIO $ getStore localCell state) >>=
flip maybe (liftIO . flip putSimpleStore state) (void $ makeHandler $ bimapEitherT show id $ EitherT $ insertStore localCell state)
-- | Handler for migrations: Write or overwrite
| 509
|
getOrInsertStore :: (SimpleCellConstraint cell k src dst tm st) => st -> DistributedHandlerM urllist st ()
getOrInsertStore state = do
localCell <- asks localCell
(liftIO $ getStore localCell state) >>=
flip maybe (liftIO . flip putSimpleStore state) (void $ makeHandler $ bimapEitherT show id $ EitherT $ insertStore localCell state)
-- | Handler for migrations: Write or overwrite
| 391
|
getOrInsertStore state = do
localCell <- asks localCell
(liftIO $ getStore localCell state) >>=
flip maybe (liftIO . flip putSimpleStore state) (void $ makeHandler $ bimapEitherT show id $ EitherT $ insertStore localCell state)
-- | Handler for migrations: Write or overwrite
| 284
| true
| true
| 0
| 13
| 90
| 125
| 61
| 64
| null | null |
gcampax/ghc
|
compiler/utils/Outputable.hs
|
bsd-3-clause
|
qualPackage :: PprStyle -> QueryQualifyPackage
qualPackage (PprUser q _) m = queryQualifyPackage q m
| 101
|
qualPackage :: PprStyle -> QueryQualifyPackage
qualPackage (PprUser q _) m = queryQualifyPackage q m
| 101
|
qualPackage (PprUser q _) m = queryQualifyPackage q m
| 54
| false
| true
| 0
| 9
| 14
| 38
| 17
| 21
| null | null |
lukexi/ghc
|
compiler/nativeGen/X86/Regs.hs
|
bsd-3-clause
|
rcx = regSingle 2
| 19
|
rcx = regSingle 2
| 19
|
rcx = regSingle 2
| 19
| false
| false
| 1
| 5
| 5
| 12
| 4
| 8
| null | null |
emc2/chill
|
src/IR/FlatIR/LLVMGen/Metadata.hs
|
bsd-3-clause
|
mutabilityValue VolatileOnce = LLVM.mdString "writeonce"
| 56
|
mutabilityValue VolatileOnce = LLVM.mdString "writeonce"
| 56
|
mutabilityValue VolatileOnce = LLVM.mdString "writeonce"
| 56
| false
| false
| 0
| 6
| 4
| 14
| 6
| 8
| null | null |
asr/fotc
|
notes/fixed-points/Functors.hs
|
mit
|
-- The stream destructors.
headS ∷ Stream a → a
headS xs = case out xs of
St x _ → x
| 97
|
headS ∷ Stream a → a
headS xs = case out xs of
St x _ → x
| 70
|
headS xs = case out xs of
St x _ → x
| 49
| true
| true
| 0
| 8
| 34
| 46
| 20
| 26
| null | null |
felixsch/moonbase-ng
|
src/Moonbase/DBus.hs
|
lgpl-2.1
|
dbusMethod :: ObjectPath -> (Ref -> [Method]) -> Moon ()
dbusMethod objPath generator = do
client <- view dbus <$> get
ref <- ask
liftIO $ export client objPath $ generator ref
| 185
|
dbusMethod :: ObjectPath -> (Ref -> [Method]) -> Moon ()
dbusMethod objPath generator = do
client <- view dbus <$> get
ref <- ask
liftIO $ export client objPath $ generator ref
| 185
|
dbusMethod objPath generator = do
client <- view dbus <$> get
ref <- ask
liftIO $ export client objPath $ generator ref
| 128
| false
| true
| 0
| 10
| 40
| 84
| 38
| 46
| null | null |
xarts19/GoHaskell
|
Main.hs
|
mit
|
keyPressed = do isKey <- c_kbhit
if isKey then Just <$> c_getch
else return Nothing
| 124
|
keyPressed = do isKey <- c_kbhit
if isKey then Just <$> c_getch
else return Nothing
| 124
|
keyPressed = do isKey <- c_kbhit
if isKey then Just <$> c_getch
else return Nothing
| 124
| false
| false
| 0
| 8
| 55
| 31
| 15
| 16
| null | null |
szehk/Haskell-Carbonara-Library
|
src/Data/Carbonara/Char.hs
|
bsd-3-clause
|
isNotOctDigit :: Char -> Bool
isNotOctDigit = not . isOctDigit
| 62
|
isNotOctDigit :: Char -> Bool
isNotOctDigit = not . isOctDigit
| 62
|
isNotOctDigit = not . isOctDigit
| 32
| false
| true
| 0
| 5
| 9
| 19
| 10
| 9
| null | null |
agocorona/ghcjs-perch
|
src/Internal/Perch.hs
|
mit
|
img = nelem "img"
| 22
|
img = nelem "img"
| 22
|
img = nelem "img"
| 22
| false
| false
| 1
| 5
| 8
| 13
| 4
| 9
| null | null |
copumpkin/vector-static
|
Data/Vector/Unboxed/Static.hs
|
bsd-3-clause
|
zip4 (Vec as) (Vec bs) (Vec cs) (Vec ds) = Vec (G.zip4 as bs cs ds)
| 67
|
zip4 (Vec as) (Vec bs) (Vec cs) (Vec ds) = Vec (G.zip4 as bs cs ds)
| 67
|
zip4 (Vec as) (Vec bs) (Vec cs) (Vec ds) = Vec (G.zip4 as bs cs ds)
| 67
| false
| false
| 0
| 8
| 15
| 56
| 27
| 29
| null | null |
kazu-yamamoto/http-types
|
Network/HTTP/Types/Status.hs
|
bsd-3-clause
|
-- | Use Proxy 305
useProxy305 :: Status
useProxy305 = status305
| 64
|
useProxy305 :: Status
useProxy305 = status305
| 45
|
useProxy305 = status305
| 23
| true
| true
| 0
| 6
| 10
| 19
| 8
| 11
| null | null |
apyrgio/snf-ganeti
|
src/Ganeti/HTools/Node.hs
|
bsd-2-clause
|
-- | Calculate the new spindle usage
calcSpindleUse ::
Bool -- Action: True = adding instance, False = removing it
-> Node -> Instance.Instance -> Double
calcSpindleUse _ (Node {exclStorage = True}) _ = 0.0
| 239
|
calcSpindleUse ::
Bool -- Action: True = adding instance, False = removing it
-> Node -> Instance.Instance -> Double
calcSpindleUse _ (Node {exclStorage = True}) _ = 0.0
| 202
|
calcSpindleUse _ (Node {exclStorage = True}) _ = 0.0
| 52
| true
| true
| 4
| 9
| 67
| 54
| 26
| 28
| null | null |
Arguggi/Frinfo
|
src/Frinfo/Config.hs
|
mit
|
downSpeedIcon :: T.Text
downSpeedIcon = "/home/arguggi/dotfiles/icons/xbm8x8/net_down_03.xbm"
| 93
|
downSpeedIcon :: T.Text
downSpeedIcon = "/home/arguggi/dotfiles/icons/xbm8x8/net_down_03.xbm"
| 93
|
downSpeedIcon = "/home/arguggi/dotfiles/icons/xbm8x8/net_down_03.xbm"
| 69
| false
| true
| 0
| 5
| 5
| 13
| 7
| 6
| null | null |
donnie4w/tim
|
protocols/gen-hs/ITim_Client.hs
|
apache-2.0
|
send_timAck op arg_ab = do
seq <- seqid
seqn <- R.readIORef seq
T.writeMessageBegin op ("timAck", T.M_ONEWAY, seqn)
write_TimAck_args op (TimAck_args{timAck_args_ab=arg_ab})
T.writeMessageEnd op
T.tFlush (T.getTransport op)
| 235
|
send_timAck op arg_ab = do
seq <- seqid
seqn <- R.readIORef seq
T.writeMessageBegin op ("timAck", T.M_ONEWAY, seqn)
write_TimAck_args op (TimAck_args{timAck_args_ab=arg_ab})
T.writeMessageEnd op
T.tFlush (T.getTransport op)
| 235
|
send_timAck op arg_ab = do
seq <- seqid
seqn <- R.readIORef seq
T.writeMessageBegin op ("timAck", T.M_ONEWAY, seqn)
write_TimAck_args op (TimAck_args{timAck_args_ab=arg_ab})
T.writeMessageEnd op
T.tFlush (T.getTransport op)
| 235
| false
| false
| 1
| 11
| 36
| 99
| 44
| 55
| null | null |
ollef/Grempa
|
Data/Parser/Grempa/Grammar.hs
|
bsd-3-clause
|
severalInter :: ( ToSym s x, ToSymT s x ~ a
, ToSym s t, ToSymT s t ~ s
, Typeable a, Typeable s)
=> t -> x -> Grammar s [a]
severalInter tok x = do
rec
xs <- rule [ (:[]) <@> x
, (:) <@> x <# tok <#> xs]
return xs
-- | Takes two symbols and combines them with @(:)@.
--
-- Creates one new rule.
--
-- This can for example be used instead of using both 'several' and 'several0'
-- on the same symbol, as that will create three new rules, whereas the
-- equivalent using 'cons' will only create two new rules. Example
-- transformation:
--
-- > xs0 <- several0 x
-- > xs <- several x
-- > ==>
-- > xs0 <- several0 x
-- > xs <- x `cons` xs0
| 724
|
severalInter :: ( ToSym s x, ToSymT s x ~ a
, ToSym s t, ToSymT s t ~ s
, Typeable a, Typeable s)
=> t -> x -> Grammar s [a]
severalInter tok x = do
rec
xs <- rule [ (:[]) <@> x
, (:) <@> x <# tok <#> xs]
return xs
-- | Takes two symbols and combines them with @(:)@.
--
-- Creates one new rule.
--
-- This can for example be used instead of using both 'several' and 'several0'
-- on the same symbol, as that will create three new rules, whereas the
-- equivalent using 'cons' will only create two new rules. Example
-- transformation:
--
-- > xs0 <- several0 x
-- > xs <- several x
-- > ==>
-- > xs0 <- several0 x
-- > xs <- x `cons` xs0
| 724
|
severalInter tok x = do
rec
xs <- rule [ (:[]) <@> x
, (:) <@> x <# tok <#> xs]
return xs
-- | Takes two symbols and combines them with @(:)@.
--
-- Creates one new rule.
--
-- This can for example be used instead of using both 'several' and 'several0'
-- on the same symbol, as that will create three new rules, whereas the
-- equivalent using 'cons' will only create two new rules. Example
-- transformation:
--
-- > xs0 <- several0 x
-- > xs <- several x
-- > ==>
-- > xs0 <- several0 x
-- > xs <- x `cons` xs0
| 554
| false
| true
| 0
| 15
| 228
| 157
| 84
| 73
| null | null |
uuhan/Idris-dev
|
src/Idris/Output.hs
|
bsd-3-clause
|
iPrintFunTypes bnd n overloads = do ist <- getIState
let ppo = ppOptionIst ist
let infixes = idris_infixes ist
let output = vsep (map (uncurry (ppOverload ppo infixes)) overloads)
iRenderResult output
where fullName ppo n | length overloads > 1 = prettyName True True bnd n
| otherwise = prettyName True (ppopt_impl ppo) bnd n
ppOverload ppo infixes n tm =
fullName ppo n <+> colon <+> align tm
| 581
|
iPrintFunTypes bnd n overloads = do ist <- getIState
let ppo = ppOptionIst ist
let infixes = idris_infixes ist
let output = vsep (map (uncurry (ppOverload ppo infixes)) overloads)
iRenderResult output
where fullName ppo n | length overloads > 1 = prettyName True True bnd n
| otherwise = prettyName True (ppopt_impl ppo) bnd n
ppOverload ppo infixes n tm =
fullName ppo n <+> colon <+> align tm
| 581
|
iPrintFunTypes bnd n overloads = do ist <- getIState
let ppo = ppOptionIst ist
let infixes = idris_infixes ist
let output = vsep (map (uncurry (ppOverload ppo infixes)) overloads)
iRenderResult output
where fullName ppo n | length overloads > 1 = prettyName True True bnd n
| otherwise = prettyName True (ppopt_impl ppo) bnd n
ppOverload ppo infixes n tm =
fullName ppo n <+> colon <+> align tm
| 581
| false
| false
| 5
| 16
| 255
| 183
| 77
| 106
| null | null |
tolysz/yesod
|
yesod-core/Yesod/Core/Content.hs
|
mit
|
typeJpeg :: ContentType
typeJpeg = "image/jpeg"
| 47
|
typeJpeg :: ContentType
typeJpeg = "image/jpeg"
| 47
|
typeJpeg = "image/jpeg"
| 23
| false
| true
| 0
| 6
| 5
| 18
| 7
| 11
| null | null |
ashishnegi/hsalgos
|
src/LongestEdgePathInDAG.hs
|
bsd-3-clause
|
isValidPath :: [NodeId] -> DAG -> (Int, Int) -> Bool
isValidPath path graph (width, height) =
let heights = fmap (heightOfNode . flip nodeData graph) path
incOrderHeights = any (\(a,b) -> a <= b) $ zip heights (drop 1 heights)
validTransitions = all (isValidTransition width) $ zip path (drop 1 path)
in incOrderHeights && validTransitions
where
isValidTransition width (nodeId1, nodeId2) =
case nodeId1 - nodeId2 of
1 -> nodeId1 `mod` width /= 0 -- left
-1 -> nodeId1 `mod` width /= (width - 1) -- right
w -> (w == -width) || (w == width) --up/down
| 601
|
isValidPath :: [NodeId] -> DAG -> (Int, Int) -> Bool
isValidPath path graph (width, height) =
let heights = fmap (heightOfNode . flip nodeData graph) path
incOrderHeights = any (\(a,b) -> a <= b) $ zip heights (drop 1 heights)
validTransitions = all (isValidTransition width) $ zip path (drop 1 path)
in incOrderHeights && validTransitions
where
isValidTransition width (nodeId1, nodeId2) =
case nodeId1 - nodeId2 of
1 -> nodeId1 `mod` width /= 0 -- left
-1 -> nodeId1 `mod` width /= (width - 1) -- right
w -> (w == -width) || (w == width) --up/down
| 601
|
isValidPath path graph (width, height) =
let heights = fmap (heightOfNode . flip nodeData graph) path
incOrderHeights = any (\(a,b) -> a <= b) $ zip heights (drop 1 heights)
validTransitions = all (isValidTransition width) $ zip path (drop 1 path)
in incOrderHeights && validTransitions
where
isValidTransition width (nodeId1, nodeId2) =
case nodeId1 - nodeId2 of
1 -> nodeId1 `mod` width /= 0 -- left
-1 -> nodeId1 `mod` width /= (width - 1) -- right
w -> (w == -width) || (w == width) --up/down
| 548
| false
| true
| 0
| 13
| 147
| 248
| 132
| 116
| null | null |
jwiegley/ghc-release
|
libraries/Cabal/cabal-install/Distribution/Client/Setup.hs
|
gpl-3.0
|
defaultMaxBackjumps :: Int
defaultMaxBackjumps = 200
| 52
|
defaultMaxBackjumps :: Int
defaultMaxBackjumps = 200
| 52
|
defaultMaxBackjumps = 200
| 25
| false
| true
| 0
| 4
| 5
| 11
| 6
| 5
| null | null |
peterokagey/haskellOEIS
|
src/External/A065880.hs
|
apache-2.0
|
a065880 n = n * a272760 n
| 25
|
a065880 n = n * a272760 n
| 25
|
a065880 n = n * a272760 n
| 25
| false
| false
| 0
| 6
| 6
| 16
| 7
| 9
| null | null |
sdiehl/ghc
|
compiler/GHC/Cmm/Info/Build.hs
|
bsd-3-clause
|
doSRTs
:: DynFlags
-> ModuleSRTInfo
-> [(CAFEnv, [CmmDecl])]
-> [(CAFSet, CmmDecl)]
-> IO (ModuleSRTInfo, [CmmDeclSRTs])
doSRTs dflags moduleSRTInfo procs data_ = do
us <- mkSplitUniqSupply 'u'
-- Ignore the original grouping of decls, and combine all the
-- CAFEnvs into a single CAFEnv.
let static_data_env :: Map CLabel CAFSet
static_data_env =
Map.fromList $
flip map data_ $
\(set, decl) ->
case decl of
CmmProc{} ->
pprPanic "doSRTs" (text "Proc in static data list:" <+> ppr decl)
CmmData _ static ->
case static of
CmmStatics lbl _ _ _ -> (lbl, set)
CmmStaticsRaw lbl _ -> (lbl, set)
static_data :: Set CLabel
static_data = Map.keysSet static_data_env
(proc_envs, procss) = unzip procs
cafEnv = mapUnions proc_envs
decls = map snd data_ ++ concat procss
staticFuns = mapFromList (getStaticFuns decls)
-- Put the decls in dependency order. Why? So that we can implement
-- [Inline] and [Filter]. If we need to refer to an SRT that has
-- a single entry, we use the entry itself, which means that we
-- don't need to generate the singleton SRT in the first place. But
-- to do this we need to process blocks before things that depend on
-- them.
let
sccs :: [SCC (SomeLabel, CAFLabel, Set CAFLabel)]
sccs = {-# SCC depAnalSRTs #-} depAnalSRTs cafEnv static_data_env decls
cafsWithSRTs :: [(Label, CAFLabel, Set CAFLabel)]
cafsWithSRTs = getCAFs cafEnv decls
srtTraceM "doSRTs" (text "data:" <+> ppr data_ $$
text "procs:" <+> ppr procs $$
text "static_data_env:" <+> ppr static_data_env $$
text "sccs:" <+> ppr sccs $$
text "cafsWithSRTs:" <+> ppr cafsWithSRTs)
-- On each strongly-connected group of decls, construct the SRT
-- closures and the SRT fields for info tables.
let result ::
[ ( [CmmDeclSRTs] -- generated SRTs
, [(Label, CLabel)] -- SRT fields for info tables
, [(Label, [SRTEntry])] -- SRTs to attach to static functions
, Bool -- Whether the group has CAF references
) ]
(result, moduleSRTInfo') =
initUs_ us $
flip runStateT moduleSRTInfo $ do
nonCAFs <- mapM (doSCC dflags staticFuns static_data) sccs
cAFs <- forM cafsWithSRTs $ \(l, cafLbl, cafs) ->
oneSRT dflags staticFuns [BlockLabel l] [cafLbl]
True{-is a CAF-} cafs static_data
return (nonCAFs ++ cAFs)
(srt_declss, pairs, funSRTs, has_caf_refs) = unzip4 result
srt_decls = concat srt_declss
unless (null srt_decls) $
dumpIfSet_dyn dflags Opt_D_dump_srts "SRTs" FormatCMM (ppr srt_decls)
-- Next, update the info tables with the SRTs
let
srtFieldMap = mapFromList (concat pairs)
funSRTMap = mapFromList (concat funSRTs)
has_caf_refs' = or has_caf_refs
decls' =
concatMap (updInfoSRTs dflags srtFieldMap funSRTMap has_caf_refs') decls
-- Finally update CafInfos for raw static literals (CmmStaticsRaw). Those are
-- not analysed in oneSRT so we never add entries for them to the SRTMap.
let srtMap_w_raws =
foldl' (\(srtMap :: SRTMap) (_, decl) ->
case decl of
CmmData _ CmmStatics{} ->
-- already updated by oneSRT
srtMap
CmmData _ (CmmStaticsRaw lbl _)
| isIdLabel lbl ->
-- not analysed by oneSRT, declare it non-CAFFY here
Map.insert (mkCAFLabel lbl) Nothing srtMap
| otherwise ->
-- Not an IdLabel, ignore
srtMap
CmmProc{} ->
pprPanic "doSRTs" (text "Found Proc in static data list:" <+> ppr decl))
(moduleSRTMap moduleSRTInfo') data_
return (moduleSRTInfo'{ moduleSRTMap = srtMap_w_raws }, srt_decls ++ decls')
-- | Build the SRT for a strongly-connected component of blocks
| 4,213
|
doSRTs
:: DynFlags
-> ModuleSRTInfo
-> [(CAFEnv, [CmmDecl])]
-> [(CAFSet, CmmDecl)]
-> IO (ModuleSRTInfo, [CmmDeclSRTs])
doSRTs dflags moduleSRTInfo procs data_ = do
us <- mkSplitUniqSupply 'u'
-- Ignore the original grouping of decls, and combine all the
-- CAFEnvs into a single CAFEnv.
let static_data_env :: Map CLabel CAFSet
static_data_env =
Map.fromList $
flip map data_ $
\(set, decl) ->
case decl of
CmmProc{} ->
pprPanic "doSRTs" (text "Proc in static data list:" <+> ppr decl)
CmmData _ static ->
case static of
CmmStatics lbl _ _ _ -> (lbl, set)
CmmStaticsRaw lbl _ -> (lbl, set)
static_data :: Set CLabel
static_data = Map.keysSet static_data_env
(proc_envs, procss) = unzip procs
cafEnv = mapUnions proc_envs
decls = map snd data_ ++ concat procss
staticFuns = mapFromList (getStaticFuns decls)
-- Put the decls in dependency order. Why? So that we can implement
-- [Inline] and [Filter]. If we need to refer to an SRT that has
-- a single entry, we use the entry itself, which means that we
-- don't need to generate the singleton SRT in the first place. But
-- to do this we need to process blocks before things that depend on
-- them.
let
sccs :: [SCC (SomeLabel, CAFLabel, Set CAFLabel)]
sccs = {-# SCC depAnalSRTs #-} depAnalSRTs cafEnv static_data_env decls
cafsWithSRTs :: [(Label, CAFLabel, Set CAFLabel)]
cafsWithSRTs = getCAFs cafEnv decls
srtTraceM "doSRTs" (text "data:" <+> ppr data_ $$
text "procs:" <+> ppr procs $$
text "static_data_env:" <+> ppr static_data_env $$
text "sccs:" <+> ppr sccs $$
text "cafsWithSRTs:" <+> ppr cafsWithSRTs)
-- On each strongly-connected group of decls, construct the SRT
-- closures and the SRT fields for info tables.
let result ::
[ ( [CmmDeclSRTs] -- generated SRTs
, [(Label, CLabel)] -- SRT fields for info tables
, [(Label, [SRTEntry])] -- SRTs to attach to static functions
, Bool -- Whether the group has CAF references
) ]
(result, moduleSRTInfo') =
initUs_ us $
flip runStateT moduleSRTInfo $ do
nonCAFs <- mapM (doSCC dflags staticFuns static_data) sccs
cAFs <- forM cafsWithSRTs $ \(l, cafLbl, cafs) ->
oneSRT dflags staticFuns [BlockLabel l] [cafLbl]
True{-is a CAF-} cafs static_data
return (nonCAFs ++ cAFs)
(srt_declss, pairs, funSRTs, has_caf_refs) = unzip4 result
srt_decls = concat srt_declss
unless (null srt_decls) $
dumpIfSet_dyn dflags Opt_D_dump_srts "SRTs" FormatCMM (ppr srt_decls)
-- Next, update the info tables with the SRTs
let
srtFieldMap = mapFromList (concat pairs)
funSRTMap = mapFromList (concat funSRTs)
has_caf_refs' = or has_caf_refs
decls' =
concatMap (updInfoSRTs dflags srtFieldMap funSRTMap has_caf_refs') decls
-- Finally update CafInfos for raw static literals (CmmStaticsRaw). Those are
-- not analysed in oneSRT so we never add entries for them to the SRTMap.
let srtMap_w_raws =
foldl' (\(srtMap :: SRTMap) (_, decl) ->
case decl of
CmmData _ CmmStatics{} ->
-- already updated by oneSRT
srtMap
CmmData _ (CmmStaticsRaw lbl _)
| isIdLabel lbl ->
-- not analysed by oneSRT, declare it non-CAFFY here
Map.insert (mkCAFLabel lbl) Nothing srtMap
| otherwise ->
-- Not an IdLabel, ignore
srtMap
CmmProc{} ->
pprPanic "doSRTs" (text "Found Proc in static data list:" <+> ppr decl))
(moduleSRTMap moduleSRTInfo') data_
return (moduleSRTInfo'{ moduleSRTMap = srtMap_w_raws }, srt_decls ++ decls')
-- | Build the SRT for a strongly-connected component of blocks
| 4,212
|
doSRTs dflags moduleSRTInfo procs data_ = do
us <- mkSplitUniqSupply 'u'
-- Ignore the original grouping of decls, and combine all the
-- CAFEnvs into a single CAFEnv.
let static_data_env :: Map CLabel CAFSet
static_data_env =
Map.fromList $
flip map data_ $
\(set, decl) ->
case decl of
CmmProc{} ->
pprPanic "doSRTs" (text "Proc in static data list:" <+> ppr decl)
CmmData _ static ->
case static of
CmmStatics lbl _ _ _ -> (lbl, set)
CmmStaticsRaw lbl _ -> (lbl, set)
static_data :: Set CLabel
static_data = Map.keysSet static_data_env
(proc_envs, procss) = unzip procs
cafEnv = mapUnions proc_envs
decls = map snd data_ ++ concat procss
staticFuns = mapFromList (getStaticFuns decls)
-- Put the decls in dependency order. Why? So that we can implement
-- [Inline] and [Filter]. If we need to refer to an SRT that has
-- a single entry, we use the entry itself, which means that we
-- don't need to generate the singleton SRT in the first place. But
-- to do this we need to process blocks before things that depend on
-- them.
let
sccs :: [SCC (SomeLabel, CAFLabel, Set CAFLabel)]
sccs = {-# SCC depAnalSRTs #-} depAnalSRTs cafEnv static_data_env decls
cafsWithSRTs :: [(Label, CAFLabel, Set CAFLabel)]
cafsWithSRTs = getCAFs cafEnv decls
srtTraceM "doSRTs" (text "data:" <+> ppr data_ $$
text "procs:" <+> ppr procs $$
text "static_data_env:" <+> ppr static_data_env $$
text "sccs:" <+> ppr sccs $$
text "cafsWithSRTs:" <+> ppr cafsWithSRTs)
-- On each strongly-connected group of decls, construct the SRT
-- closures and the SRT fields for info tables.
let result ::
[ ( [CmmDeclSRTs] -- generated SRTs
, [(Label, CLabel)] -- SRT fields for info tables
, [(Label, [SRTEntry])] -- SRTs to attach to static functions
, Bool -- Whether the group has CAF references
) ]
(result, moduleSRTInfo') =
initUs_ us $
flip runStateT moduleSRTInfo $ do
nonCAFs <- mapM (doSCC dflags staticFuns static_data) sccs
cAFs <- forM cafsWithSRTs $ \(l, cafLbl, cafs) ->
oneSRT dflags staticFuns [BlockLabel l] [cafLbl]
True{-is a CAF-} cafs static_data
return (nonCAFs ++ cAFs)
(srt_declss, pairs, funSRTs, has_caf_refs) = unzip4 result
srt_decls = concat srt_declss
unless (null srt_decls) $
dumpIfSet_dyn dflags Opt_D_dump_srts "SRTs" FormatCMM (ppr srt_decls)
-- Next, update the info tables with the SRTs
let
srtFieldMap = mapFromList (concat pairs)
funSRTMap = mapFromList (concat funSRTs)
has_caf_refs' = or has_caf_refs
decls' =
concatMap (updInfoSRTs dflags srtFieldMap funSRTMap has_caf_refs') decls
-- Finally update CafInfos for raw static literals (CmmStaticsRaw). Those are
-- not analysed in oneSRT so we never add entries for them to the SRTMap.
let srtMap_w_raws =
foldl' (\(srtMap :: SRTMap) (_, decl) ->
case decl of
CmmData _ CmmStatics{} ->
-- already updated by oneSRT
srtMap
CmmData _ (CmmStaticsRaw lbl _)
| isIdLabel lbl ->
-- not analysed by oneSRT, declare it non-CAFFY here
Map.insert (mkCAFLabel lbl) Nothing srtMap
| otherwise ->
-- Not an IdLabel, ignore
srtMap
CmmProc{} ->
pprPanic "doSRTs" (text "Found Proc in static data list:" <+> ppr decl))
(moduleSRTMap moduleSRTInfo') data_
return (moduleSRTInfo'{ moduleSRTMap = srtMap_w_raws }, srt_decls ++ decls')
-- | Build the SRT for a strongly-connected component of blocks
| 4,081
| false
| true
| 0
| 20
| 1,382
| 910
| 468
| 442
| null | null |
anttisalonen/nix
|
src/Filter.hs
|
gpl-3.0
|
filterAll (f:fs) xs = filterAll fs (filter f xs)
| 48
|
filterAll (f:fs) xs = filterAll fs (filter f xs)
| 48
|
filterAll (f:fs) xs = filterAll fs (filter f xs)
| 48
| false
| false
| 0
| 7
| 8
| 31
| 15
| 16
| null | null |
acowley/ghc
|
compiler/prelude/PrelNames.hs
|
bsd-3-clause
|
ordClassKey = mkPreludeClassUnique 12
| 49
|
ordClassKey = mkPreludeClassUnique 12
| 49
|
ordClassKey = mkPreludeClassUnique 12
| 49
| false
| false
| 0
| 5
| 15
| 9
| 4
| 5
| null | null |
fumieval/audiovisual
|
src/Data/Graphics/Bitmap.hs
|
bsd-3-clause
|
bbox :: Bitmap -> B.Box V2 Int
bbox (Bitmap (C.Image w h _) (V2 x y) _) = B.Box (V2 x y) (V2 (x+w) (y+h))
| 105
|
bbox :: Bitmap -> B.Box V2 Int
bbox (Bitmap (C.Image w h _) (V2 x y) _) = B.Box (V2 x y) (V2 (x+w) (y+h))
| 105
|
bbox (Bitmap (C.Image w h _) (V2 x y) _) = B.Box (V2 x y) (V2 (x+w) (y+h))
| 74
| false
| true
| 0
| 9
| 24
| 96
| 47
| 49
| null | null |
gazay/faker
|
src/Faker/Business.hs
|
mit
|
-- | Returns random business-like credit card number, i.e. "1234-2121-1221-1211"
creditCardNumber :: Faker String
creditCardNumber = randomBusinessWord "credit_card_number"
| 172
|
creditCardNumber :: Faker String
creditCardNumber = randomBusinessWord "credit_card_number"
| 91
|
creditCardNumber = randomBusinessWord "credit_card_number"
| 58
| true
| true
| 0
| 5
| 17
| 18
| 9
| 9
| null | null |
guilt/webify
|
src/Utils.hs
|
mit
|
toStrict :: BL.ByteString -> B.ByteString
toStrict = B.concat . BL.toChunks
| 75
|
toStrict :: BL.ByteString -> B.ByteString
toStrict = B.concat . BL.toChunks
| 75
|
toStrict = B.concat . BL.toChunks
| 33
| false
| true
| 0
| 8
| 9
| 34
| 15
| 19
| null | null |
codeq/language-py
|
src/Language/Py/LexerUtils.hs
|
bsd-3-clause
|
utf8Encode :: Char -> [Word8]
utf8Encode = map fromIntegral . go . ord
where
go oc
| oc <= 0x7f = [oc]
| oc <= 0x7ff = [ 0xc0 + (oc `Data.Bits.shiftR` 6)
, 0x80 + oc Data.Bits..&. 0x3f
]
| oc <= 0xffff = [ 0xe0 + (oc `Data.Bits.shiftR` 12)
, 0x80 + ((oc `Data.Bits.shiftR` 6) Data.Bits..&. 0x3f)
, 0x80 + oc Data.Bits..&. 0x3f
]
| otherwise = [ 0xf0 + (oc `Data.Bits.shiftR` 18)
, 0x80 + ((oc `Data.Bits.shiftR` 12) Data.Bits..&. 0x3f)
, 0x80 + ((oc `Data.Bits.shiftR` 6) Data.Bits..&. 0x3f)
, 0x80 + oc Data.Bits..&. 0x3f
]
| 821
|
utf8Encode :: Char -> [Word8]
utf8Encode = map fromIntegral . go . ord
where
go oc
| oc <= 0x7f = [oc]
| oc <= 0x7ff = [ 0xc0 + (oc `Data.Bits.shiftR` 6)
, 0x80 + oc Data.Bits..&. 0x3f
]
| oc <= 0xffff = [ 0xe0 + (oc `Data.Bits.shiftR` 12)
, 0x80 + ((oc `Data.Bits.shiftR` 6) Data.Bits..&. 0x3f)
, 0x80 + oc Data.Bits..&. 0x3f
]
| otherwise = [ 0xf0 + (oc `Data.Bits.shiftR` 18)
, 0x80 + ((oc `Data.Bits.shiftR` 12) Data.Bits..&. 0x3f)
, 0x80 + ((oc `Data.Bits.shiftR` 6) Data.Bits..&. 0x3f)
, 0x80 + oc Data.Bits..&. 0x3f
]
| 821
|
utf8Encode = map fromIntegral . go . ord
where
go oc
| oc <= 0x7f = [oc]
| oc <= 0x7ff = [ 0xc0 + (oc `Data.Bits.shiftR` 6)
, 0x80 + oc Data.Bits..&. 0x3f
]
| oc <= 0xffff = [ 0xe0 + (oc `Data.Bits.shiftR` 12)
, 0x80 + ((oc `Data.Bits.shiftR` 6) Data.Bits..&. 0x3f)
, 0x80 + oc Data.Bits..&. 0x3f
]
| otherwise = [ 0xf0 + (oc `Data.Bits.shiftR` 18)
, 0x80 + ((oc `Data.Bits.shiftR` 12) Data.Bits..&. 0x3f)
, 0x80 + ((oc `Data.Bits.shiftR` 6) Data.Bits..&. 0x3f)
, 0x80 + oc Data.Bits..&. 0x3f
]
| 791
| false
| true
| 0
| 11
| 394
| 273
| 156
| 117
| null | null |
adinapoli/Shelly.hs
|
src/Shelly.hs
|
bsd-3-clause
|
traceCanonicPath :: (Text -> Text) -> FilePath -> Sh FilePath
traceCanonicPath = tracePath canonic
| 98
|
traceCanonicPath :: (Text -> Text) -> FilePath -> Sh FilePath
traceCanonicPath = tracePath canonic
| 98
|
traceCanonicPath = tracePath canonic
| 36
| false
| true
| 0
| 7
| 13
| 32
| 16
| 16
| null | null |
ssaavedra/liquidhaskell
|
tests/pos/StructRec.hs
|
bsd-3-clause
|
insert y (Ln x xs)
| y <= x = Ln y (Ln x xs)
| otherwise = Ln x (insert y xs)
| 94
|
insert y (Ln x xs)
| y <= x = Ln y (Ln x xs)
| otherwise = Ln x (insert y xs)
| 94
|
insert y (Ln x xs)
| y <= x = Ln y (Ln x xs)
| otherwise = Ln x (insert y xs)
| 94
| false
| false
| 0
| 8
| 39
| 68
| 29
| 39
| null | null |
jcpetruzza/haskell-ast
|
src/Language/Haskell/AST/HSE.hs
|
bsd-3-clause
|
fromHseActivation (E.ActiveUntil l k) = Pragmas.ActiveUntil l k
| 64
|
fromHseActivation (E.ActiveUntil l k) = Pragmas.ActiveUntil l k
| 64
|
fromHseActivation (E.ActiveUntil l k) = Pragmas.ActiveUntil l k
| 64
| false
| false
| 0
| 7
| 8
| 27
| 12
| 15
| null | null |
damianfral/clay
|
src/Clay/Font.hs
|
bsd-3-clause
|
monospace = GenericFontFamily "monospace"
| 41
|
monospace = GenericFontFamily "monospace"
| 41
|
monospace = GenericFontFamily "monospace"
| 41
| false
| false
| 0
| 5
| 3
| 9
| 4
| 5
| null | null |
Teino1978-Corp/Teino1978-Corp-alex
|
src/NFA.hs
|
bsd-3-clause
|
bytesEdge _ _ _ _ = undefined
| 29
|
bytesEdge _ _ _ _ = undefined
| 29
|
bytesEdge _ _ _ _ = undefined
| 29
| false
| false
| 0
| 5
| 6
| 15
| 7
| 8
| null | null |
dorchard/camfort
|
tests/Camfort/Specification/StencilsSpec.hs
|
apache-2.0
|
-- Indices for the 2D five point stencil (deliberately in an odd order)
fivepoint = [ Cons (-1) (Cons 0 Nil), Cons 0 (Cons (-1) Nil)
, Cons 1 (Cons 0 Nil) , Cons 0 (Cons 1 Nil), Cons 0 (Cons 0 Nil)
]
| 223
|
fivepoint = [ Cons (-1) (Cons 0 Nil), Cons 0 (Cons (-1) Nil)
, Cons 1 (Cons 0 Nil) , Cons 0 (Cons 1 Nil), Cons 0 (Cons 0 Nil)
]
| 151
|
fivepoint = [ Cons (-1) (Cons 0 Nil), Cons 0 (Cons (-1) Nil)
, Cons 1 (Cons 0 Nil) , Cons 0 (Cons 1 Nil), Cons 0 (Cons 0 Nil)
]
| 151
| true
| false
| 0
| 10
| 67
| 97
| 50
| 47
| null | null |
dmjio/CompactMap
|
src/Data/CompactMap/Index.hs
|
bsd-3-clause
|
newKeyCursor :: Buffer -> Lazy.ByteString -> IO (Ptr KeyCursor)
newKeyCursor buffer keyE
= withBytes buffer (intSize*2 + keyLen) $ \keyPtr ->
do poke (castPtr keyPtr) nullPtr
putByteString (keyPtr `plusPtr` intSize) keyE keyLen
return keyPtr
where keyLen = fromIntegral $ Lazy.length keyE
| 320
|
newKeyCursor :: Buffer -> Lazy.ByteString -> IO (Ptr KeyCursor)
newKeyCursor buffer keyE
= withBytes buffer (intSize*2 + keyLen) $ \keyPtr ->
do poke (castPtr keyPtr) nullPtr
putByteString (keyPtr `plusPtr` intSize) keyE keyLen
return keyPtr
where keyLen = fromIntegral $ Lazy.length keyE
| 320
|
newKeyCursor buffer keyE
= withBytes buffer (intSize*2 + keyLen) $ \keyPtr ->
do poke (castPtr keyPtr) nullPtr
putByteString (keyPtr `plusPtr` intSize) keyE keyLen
return keyPtr
where keyLen = fromIntegral $ Lazy.length keyE
| 256
| false
| true
| 3
| 11
| 72
| 125
| 57
| 68
| null | null |
anand-singh/either
|
src/Data/Either/Validation.hs
|
bsd-3-clause
|
validationToEither :: Validation e a -> Either e a
validationToEither x = case x of
Failure e -> Left e
Success a -> Right a
| 129
|
validationToEither :: Validation e a -> Either e a
validationToEither x = case x of
Failure e -> Left e
Success a -> Right a
| 129
|
validationToEither x = case x of
Failure e -> Left e
Success a -> Right a
| 78
| false
| true
| 0
| 8
| 29
| 56
| 25
| 31
| null | null |
ptitfred/ftv-vods
|
src/YouTube/Services.hs
|
bsd-3-clause
|
listVideos :: [YouTubeId] -> Client Videos
listVideos vIds = toList <$> listVideosBatch (take 50 vIds) (drop 50 vIds)
| 117
|
listVideos :: [YouTubeId] -> Client Videos
listVideos vIds = toList <$> listVideosBatch (take 50 vIds) (drop 50 vIds)
| 117
|
listVideos vIds = toList <$> listVideosBatch (take 50 vIds) (drop 50 vIds)
| 74
| false
| true
| 0
| 8
| 17
| 49
| 24
| 25
| null | null |
holzensp/ghc
|
compiler/cmm/CLabel.hs
|
bsd-3-clause
|
labelType (SRTLabel _) = DataLabel
| 59
|
labelType (SRTLabel _) = DataLabel
| 59
|
labelType (SRTLabel _) = DataLabel
| 59
| false
| false
| 0
| 6
| 29
| 16
| 7
| 9
| null | null |
WraithM/CoreCompiler
|
src/Core/LambdaLift.hs
|
bsd-3-clause
|
collectCsE (Constr i j) = ([], Constr i j)
| 42
|
collectCsE (Constr i j) = ([], Constr i j)
| 42
|
collectCsE (Constr i j) = ([], Constr i j)
| 42
| false
| false
| 0
| 7
| 8
| 30
| 15
| 15
| null | null |
thalerjonathan/phd
|
coding/learning/haskell/grahambook/Code_Solutions/tictactoe.hs
|
gpl-3.0
|
won :: Grid -> Bool
won g = wins O g || wins X g
| 48
|
won :: Grid -> Bool
won g = wins O g || wins X g
| 48
|
won g = wins O g || wins X g
| 28
| false
| true
| 0
| 6
| 14
| 35
| 16
| 19
| null | null |
romanb/amazonka
|
amazonka-redshift/gen/Network/AWS/Redshift/DescribeClusterSnapshots.hs
|
mpl-2.0
|
-- | The type of snapshots for which you are requesting information. By default,
-- snapshots of all types are returned.
--
-- Valid Values: 'automated' | 'manual'
dcs1SnapshotType :: Lens' DescribeClusterSnapshots (Maybe Text)
dcs1SnapshotType = lens _dcs1SnapshotType (\s a -> s { _dcs1SnapshotType = a })
| 307
|
dcs1SnapshotType :: Lens' DescribeClusterSnapshots (Maybe Text)
dcs1SnapshotType = lens _dcs1SnapshotType (\s a -> s { _dcs1SnapshotType = a })
| 143
|
dcs1SnapshotType = lens _dcs1SnapshotType (\s a -> s { _dcs1SnapshotType = a })
| 79
| true
| true
| 0
| 9
| 46
| 49
| 28
| 21
| null | null |
lucasdicioccio/haskell-paris-src
|
app.hs
|
apache-2.0
|
deleteAndRedirect db path model oId = do
act <- liftIO $ db $ delete' model oId
liftIO $ print act
either (\_ -> raise "could not delete from DB") (\_ -> redirect path) act
-- Like MongoDB's insert but on a BDoc instance
| 259
|
deleteAndRedirect db path model oId = do
act <- liftIO $ db $ delete' model oId
liftIO $ print act
either (\_ -> raise "could not delete from DB") (\_ -> redirect path) act
-- Like MongoDB's insert but on a BDoc instance
| 259
|
deleteAndRedirect db path model oId = do
act <- liftIO $ db $ delete' model oId
liftIO $ print act
either (\_ -> raise "could not delete from DB") (\_ -> redirect path) act
-- Like MongoDB's insert but on a BDoc instance
| 259
| false
| false
| 0
| 10
| 81
| 78
| 37
| 41
| null | null |
iblumenfeld/saw-core
|
src/Verifier/SAW/Typechecker/Context.hs
|
bsd-3-clause
|
ppTCTermGen d _ (TCLocalDef i) | 0 <= i && i < length d = d !! i
| otherwise = text $ "Bad local var index " ++ show i
| 149
|
ppTCTermGen d _ (TCLocalDef i) | 0 <= i && i < length d = d !! i
| otherwise = text $ "Bad local var index " ++ show i
| 149
|
ppTCTermGen d _ (TCLocalDef i) | 0 <= i && i < length d = d !! i
| otherwise = text $ "Bad local var index " ++ show i
| 149
| false
| false
| 0
| 10
| 61
| 64
| 29
| 35
| null | null |
phylake/avm3
|
vm/store.hs
|
mit
|
xform_opCode {- 0x5F -} i u d s m t (Abc.FindDef) = [FindDef]
| 61
|
xform_opCode {- 0x5F -} i u d s m t (Abc.FindDef) = [FindDef]
| 61
|
xform_opCode {- 0x5F -} i u d s m t (Abc.FindDef) = [FindDef]
| 61
| false
| false
| 0
| 7
| 12
| 30
| 16
| 14
| null | null |
yiannist/ganeti
|
src/Ganeti/Constants.hs
|
bsd-2-clause
|
sshsSshHostKey :: String
sshsSshHostKey = "ssh_host_key"
| 56
|
sshsSshHostKey :: String
sshsSshHostKey = "ssh_host_key"
| 56
|
sshsSshHostKey = "ssh_host_key"
| 31
| false
| true
| 0
| 4
| 5
| 11
| 6
| 5
| null | null |
saturday06/FrameworkBenchmarks
|
frameworks/Haskell/yesod/yesod-postgres/src/Main.hs
|
bsd-3-clause
|
getFortunesR :: Handler Html
getFortunesR = do
fortunesFromDb <- runPg $ selectList [] []
let fortunes = sortBy (compare `on` fortuneMessage . entityVal) $ (Entity (toSqlKey 0) Fortune{fortuneMessage="Additional fortune added at request time."}):fortunesFromDb
defaultLayout $ do
setTitle "Fortunes"
[whamlet|
<table>
<tr>
<th>id
<th>message
$forall fortune <- fortunes
<tr>
<td>#{entityKey fortune}
<td>#{fortuneMessage $ entityVal fortune}
|]
| 622
|
getFortunesR :: Handler Html
getFortunesR = do
fortunesFromDb <- runPg $ selectList [] []
let fortunes = sortBy (compare `on` fortuneMessage . entityVal) $ (Entity (toSqlKey 0) Fortune{fortuneMessage="Additional fortune added at request time."}):fortunesFromDb
defaultLayout $ do
setTitle "Fortunes"
[whamlet|
<table>
<tr>
<th>id
<th>message
$forall fortune <- fortunes
<tr>
<td>#{entityKey fortune}
<td>#{fortuneMessage $ entityVal fortune}
|]
| 622
|
getFortunesR = do
fortunesFromDb <- runPg $ selectList [] []
let fortunes = sortBy (compare `on` fortuneMessage . entityVal) $ (Entity (toSqlKey 0) Fortune{fortuneMessage="Additional fortune added at request time."}):fortunesFromDb
defaultLayout $ do
setTitle "Fortunes"
[whamlet|
<table>
<tr>
<th>id
<th>message
$forall fortune <- fortunes
<tr>
<td>#{entityKey fortune}
<td>#{fortuneMessage $ entityVal fortune}
|]
| 593
| false
| true
| 0
| 15
| 232
| 115
| 56
| 59
| null | null |
CindyLinz/Haskell-HVG
|
src/HVG/ContextState.hs
|
bsd-3-clause
|
setTextAlign :: TextAlign -> Builder info ContextState draw ()
setTextAlign val = modifyStructState $ \ctx -> ctx{ctxTextAlign = val}
| 133
|
setTextAlign :: TextAlign -> Builder info ContextState draw ()
setTextAlign val = modifyStructState $ \ctx -> ctx{ctxTextAlign = val}
| 133
|
setTextAlign val = modifyStructState $ \ctx -> ctx{ctxTextAlign = val}
| 70
| false
| true
| 0
| 8
| 18
| 46
| 24
| 22
| null | null |
mainland/nikola
|
src/Data/Vector/CUDA/UnboxedForeign.hs
|
bsd-3-clause
|
map = G.map
| 11
|
map = G.map
| 11
|
map = G.map
| 11
| false
| false
| 1
| 6
| 2
| 12
| 4
| 8
| null | null |
Ming-Tang/FP15
|
tests/ArbitraryTokens.hs
|
mit
|
a <:- b = a <:> pure b
| 22
|
a <:- b = a <:> pure b
| 22
|
a <:- b = a <:> pure b
| 22
| false
| false
| 2
| 6
| 7
| 21
| 9
| 12
| null | null |
eijian/deeplearning
|
CNN/ActLayer.hs
|
bsd-3-clause
|
{- |
deactivate
IN : activation function
image
difference from previous layer
OUT: difference and updated layer
>>> fst $ deactivate relu [fromLists [[1.5,(-2.0)]]] [fromLists [[0.5,0.1]]]
[(1><2)
[ 0.5, 0.0 ]]
-}
m0 :: Matrix R
m0 = fromLists [[0.0]]
| 275
|
m0 :: Matrix R
m0 = fromLists [[0.0]]
| 37
|
m0 = fromLists [[0.0]]
| 22
| true
| true
| 0
| 6
| 63
| 29
| 14
| 15
| null | null |
kmilner/tamarin-prover
|
lib/theory/src/Theory/Tools/Wellformedness.hs
|
gpl-3.0
|
-- | Lower-case a string.
lowerCase :: String -> String
lowerCase = map toLower
| 79
|
lowerCase :: String -> String
lowerCase = map toLower
| 53
|
lowerCase = map toLower
| 23
| true
| true
| 0
| 5
| 13
| 19
| 10
| 9
| null | null |
grandpascorpion/canon
|
Math/NumberTheory/Canon/AurifCyclo.hs
|
gpl-3.0
|
aurCandDec :: Integer -> Integer -> Bool -> Maybe (Integer, Integer)
aurCandDec xi yi b = f (fst $ crFromI xi) (fst $ crFromI yi)
where f xp yp = aurCandDecI x y n (fst $ crFromI n) b
where n = gcd (crMaxRoot $ crAbs x) (crMaxRoot $ crAbs y)
gxy = crGCD xp yp
(x, y) = (crDivStrict xp gxy, crDivStrict yp gxy) -- this will fix the input to be relatively prime
-- toDo: incorporate the factorization status when determining the correct gcd.
-- Don't call this I(nternal) function directly. The function assumes that x and y are relatively prime. Currently uses Brent logic only.
| 736
|
aurCandDec :: Integer -> Integer -> Bool -> Maybe (Integer, Integer)
aurCandDec xi yi b = f (fst $ crFromI xi) (fst $ crFromI yi)
where f xp yp = aurCandDecI x y n (fst $ crFromI n) b
where n = gcd (crMaxRoot $ crAbs x) (crMaxRoot $ crAbs y)
gxy = crGCD xp yp
(x, y) = (crDivStrict xp gxy, crDivStrict yp gxy) -- this will fix the input to be relatively prime
-- toDo: incorporate the factorization status when determining the correct gcd.
-- Don't call this I(nternal) function directly. The function assumes that x and y are relatively prime. Currently uses Brent logic only.
| 736
|
aurCandDec xi yi b = f (fst $ crFromI xi) (fst $ crFromI yi)
where f xp yp = aurCandDecI x y n (fst $ crFromI n) b
where n = gcd (crMaxRoot $ crAbs x) (crMaxRoot $ crAbs y)
gxy = crGCD xp yp
(x, y) = (crDivStrict xp gxy, crDivStrict yp gxy) -- this will fix the input to be relatively prime
-- toDo: incorporate the factorization status when determining the correct gcd.
-- Don't call this I(nternal) function directly. The function assumes that x and y are relatively prime. Currently uses Brent logic only.
| 667
| false
| true
| 0
| 10
| 266
| 176
| 88
| 88
| null | null |
DavidAlphaFox/sblog
|
src/App/Types.hs
|
bsd-3-clause
|
setTplLayout :: FilePath -> Response ()
setTplLayout layout = do
lift $ modify $ (\s -> s {tplLayout = layout} )
| 115
|
setTplLayout :: FilePath -> Response ()
setTplLayout layout = do
lift $ modify $ (\s -> s {tplLayout = layout} )
| 115
|
setTplLayout layout = do
lift $ modify $ (\s -> s {tplLayout = layout} )
| 75
| false
| true
| 0
| 11
| 23
| 50
| 26
| 24
| null | null |
diku-dk/futhark
|
src/Futhark/CodeGen/Backends/MulticoreWASM.hs
|
isc
|
fRepMyRep :: Imp.Definitions Imp.Multicore -> [JSEntryPoint]
fRepMyRep prog =
let Imp.Functions fs = Imp.defFuns prog
function (Imp.Function entry _ _ _ res args) = do
n <- entry
Just $
JSEntryPoint
{ name = nameToString n,
parameters = map (extToString . snd) args,
ret = map extToString res
}
in mapMaybe (function . snd) fs
| 415
|
fRepMyRep :: Imp.Definitions Imp.Multicore -> [JSEntryPoint]
fRepMyRep prog =
let Imp.Functions fs = Imp.defFuns prog
function (Imp.Function entry _ _ _ res args) = do
n <- entry
Just $
JSEntryPoint
{ name = nameToString n,
parameters = map (extToString . snd) args,
ret = map extToString res
}
in mapMaybe (function . snd) fs
| 415
|
fRepMyRep prog =
let Imp.Functions fs = Imp.defFuns prog
function (Imp.Function entry _ _ _ res args) = do
n <- entry
Just $
JSEntryPoint
{ name = nameToString n,
parameters = map (extToString . snd) args,
ret = map extToString res
}
in mapMaybe (function . snd) fs
| 354
| false
| true
| 0
| 16
| 143
| 149
| 71
| 78
| null | null |
stumped2/school
|
CS381/hw1/Hw1.hs
|
apache-2.0
|
ppBread :: Bread -> String
ppBread Rye = "rye"
| 46
|
ppBread :: Bread -> String
ppBread Rye = "rye"
| 46
|
ppBread Rye = "rye"
| 19
| false
| true
| 0
| 7
| 8
| 24
| 10
| 14
| null | null |
onponomarev/ganeti
|
src/Ganeti/Constants.hs
|
bsd-2-clause
|
osScriptExport :: String
osScriptExport = "export"
| 50
|
osScriptExport :: String
osScriptExport = "export"
| 50
|
osScriptExport = "export"
| 25
| false
| true
| 0
| 6
| 5
| 18
| 7
| 11
| null | null |
piyush-kurur/yesod
|
yesod/main.hs
|
mit
|
main :: IO ()
main = do
o <- execParser =<< injectDefaults "yesod" [ ("yesod.devel.extracabalarg" , \o args -> o { optCommand =
case optCommand o of
d@Devel{} -> d { develExtraArgs = args }
c -> c
})
, ("yesod.devel.ignore" , \o args -> o { optCommand =
case optCommand o of
d@Devel{} -> d { develIgnore = args }
c -> c
})
, ("yesod.build.extracabalarg" , \o args -> o { optCommand =
case optCommand o of
b@Build{} -> b { buildExtraArgs = args }
c -> c
})
] optParser'
let cabal xs = rawSystem' (cabalCommand o) xs
case optCommand o of
Init -> scaffold
Configure -> cabal ["configure"]
Build es -> touch' >> cabal ("build":es)
Touch -> touch'
Devel da s f r b _ig es -> devel (DevelOpts (optCabalPgm o == CabalDev) da (optVerbose o) r s f b) es
Keter noRebuild -> keter (cabalCommand o) noRebuild
Version -> do putStrLn ("yesod-core version:" ++ yesodVersion)
putStrLn ("yesod version:" ++ showVersion Paths_yesod.version)
AddHandler -> addHandler
Test -> do touch'
cabal ["configure", "--enable-tests", "-flibrary-only"]
cabal ["build"]
cabal ["test"]
| 2,142
|
main :: IO ()
main = do
o <- execParser =<< injectDefaults "yesod" [ ("yesod.devel.extracabalarg" , \o args -> o { optCommand =
case optCommand o of
d@Devel{} -> d { develExtraArgs = args }
c -> c
})
, ("yesod.devel.ignore" , \o args -> o { optCommand =
case optCommand o of
d@Devel{} -> d { develIgnore = args }
c -> c
})
, ("yesod.build.extracabalarg" , \o args -> o { optCommand =
case optCommand o of
b@Build{} -> b { buildExtraArgs = args }
c -> c
})
] optParser'
let cabal xs = rawSystem' (cabalCommand o) xs
case optCommand o of
Init -> scaffold
Configure -> cabal ["configure"]
Build es -> touch' >> cabal ("build":es)
Touch -> touch'
Devel da s f r b _ig es -> devel (DevelOpts (optCabalPgm o == CabalDev) da (optVerbose o) r s f b) es
Keter noRebuild -> keter (cabalCommand o) noRebuild
Version -> do putStrLn ("yesod-core version:" ++ yesodVersion)
putStrLn ("yesod version:" ++ showVersion Paths_yesod.version)
AddHandler -> addHandler
Test -> do touch'
cabal ["configure", "--enable-tests", "-flibrary-only"]
cabal ["build"]
cabal ["test"]
| 2,142
|
main = do
o <- execParser =<< injectDefaults "yesod" [ ("yesod.devel.extracabalarg" , \o args -> o { optCommand =
case optCommand o of
d@Devel{} -> d { develExtraArgs = args }
c -> c
})
, ("yesod.devel.ignore" , \o args -> o { optCommand =
case optCommand o of
d@Devel{} -> d { develIgnore = args }
c -> c
})
, ("yesod.build.extracabalarg" , \o args -> o { optCommand =
case optCommand o of
b@Build{} -> b { buildExtraArgs = args }
c -> c
})
] optParser'
let cabal xs = rawSystem' (cabalCommand o) xs
case optCommand o of
Init -> scaffold
Configure -> cabal ["configure"]
Build es -> touch' >> cabal ("build":es)
Touch -> touch'
Devel da s f r b _ig es -> devel (DevelOpts (optCabalPgm o == CabalDev) da (optVerbose o) r s f b) es
Keter noRebuild -> keter (cabalCommand o) noRebuild
Version -> do putStrLn ("yesod-core version:" ++ yesodVersion)
putStrLn ("yesod version:" ++ showVersion Paths_yesod.version)
AddHandler -> addHandler
Test -> do touch'
cabal ["configure", "--enable-tests", "-flibrary-only"]
cabal ["build"]
cabal ["test"]
| 2,128
| false
| true
| 0
| 20
| 1,270
| 460
| 231
| 229
| null | null |
bergmark/mmdoc
|
src/Types.hs
|
bsd-3-clause
|
protectImport :: Protection -> Import -> Import
protectImport p (Import _ a b c) = Import (Just p) a b c
| 104
|
protectImport :: Protection -> Import -> Import
protectImport p (Import _ a b c) = Import (Just p) a b c
| 104
|
protectImport p (Import _ a b c) = Import (Just p) a b c
| 56
| false
| true
| 0
| 7
| 20
| 56
| 26
| 30
| null | null |
nushio3/ghc
|
compiler/stranal/DmdAnal.hs
|
bsd-3-clause
|
updSigEnv :: AnalEnv -> SigEnv -> AnalEnv
updSigEnv env sigs = env { ae_sigs = sigs }
| 85
|
updSigEnv :: AnalEnv -> SigEnv -> AnalEnv
updSigEnv env sigs = env { ae_sigs = sigs }
| 85
|
updSigEnv env sigs = env { ae_sigs = sigs }
| 43
| false
| true
| 0
| 8
| 16
| 38
| 18
| 20
| null | null |
MarcusVoelker/MuCurse
|
Function.hs
|
mit
|
resolveFunction defs (MRek g) = MRek (resolveFunction defs g)
| 61
|
resolveFunction defs (MRek g) = MRek (resolveFunction defs g)
| 61
|
resolveFunction defs (MRek g) = MRek (resolveFunction defs g)
| 61
| false
| false
| 0
| 7
| 8
| 28
| 13
| 15
| null | null |
HJvT/GeBoP
|
Game.hs
|
bsd-3-clause
|
step (i:is) t
| filled t = let u = step is (children t ! i)
in Game.update i $ t {children = children t // [(i, u)]}
| otherwise = grow t
| 160
|
step (i:is) t
| filled t = let u = step is (children t ! i)
in Game.update i $ t {children = children t // [(i, u)]}
| otherwise = grow t
| 160
|
step (i:is) t
| filled t = let u = step is (children t ! i)
in Game.update i $ t {children = children t // [(i, u)]}
| otherwise = grow t
| 160
| false
| false
| 0
| 13
| 55
| 104
| 49
| 55
| null | null |
psfblair/freepalace
|
src/FreePalace/Handlers/Incoming.hs
|
apache-2.0
|
handleInboundEvent clientState (InboundMessages.UserListMessage userListing) = handleUserList clientState userListing
| 117
|
handleInboundEvent clientState (InboundMessages.UserListMessage userListing) = handleUserList clientState userListing
| 117
|
handleInboundEvent clientState (InboundMessages.UserListMessage userListing) = handleUserList clientState userListing
| 117
| false
| false
| 0
| 8
| 7
| 24
| 11
| 13
| null | null |
jaapweel/piffle
|
src/Syntax/PPrint.hs
|
gpl-2.0
|
isEmpty _ = False
| 21
|
isEmpty _ = False
| 21
|
isEmpty _ = False
| 21
| false
| false
| 0
| 5
| 7
| 9
| 4
| 5
| null | null |
keithodulaigh/Hets
|
Comorphisms/CASL2TopSort.hs
|
gpl-2.0
|
genDisjunction :: [VAR_DECL] -> Set.Set [Maybe PRED_NAME] -> Maybe (FORMULA f)
genDisjunction vars spn
| Set.size spn == 1 =
case disjs of
[] -> Nothing
[x] -> Just x
_ -> error "CASL2TopSort.genDisjunction: this cannot happen"
| null disjs = Nothing
| otherwise = Just (disjunct disjs)
where disjs = foldl genConjunction [] (Set.toList spn)
genConjunction acc pns
| null conjs = acc
| otherwise = conjunct (reverse conjs) : acc
where conjs = foldl genPred [] (zip vars pns)
genPred acc (v, mpn) = maybe acc
(\ pn -> genPredication pn [v] : acc) mpn
{- | Each membership test of a subsort is transformed to a predication
of the corresponding unary predicate. Variables quantified over a
subsort yield a premise to the quantified formula that the
corresponding predicate holds. All typings are adjusted according
to the subsortmap and sort generation constraints are translated to
disjointness axioms. -}
| 1,042
|
genDisjunction :: [VAR_DECL] -> Set.Set [Maybe PRED_NAME] -> Maybe (FORMULA f)
genDisjunction vars spn
| Set.size spn == 1 =
case disjs of
[] -> Nothing
[x] -> Just x
_ -> error "CASL2TopSort.genDisjunction: this cannot happen"
| null disjs = Nothing
| otherwise = Just (disjunct disjs)
where disjs = foldl genConjunction [] (Set.toList spn)
genConjunction acc pns
| null conjs = acc
| otherwise = conjunct (reverse conjs) : acc
where conjs = foldl genPred [] (zip vars pns)
genPred acc (v, mpn) = maybe acc
(\ pn -> genPredication pn [v] : acc) mpn
{- | Each membership test of a subsort is transformed to a predication
of the corresponding unary predicate. Variables quantified over a
subsort yield a premise to the quantified formula that the
corresponding predicate holds. All typings are adjusted according
to the subsortmap and sort generation constraints are translated to
disjointness axioms. -}
| 1,042
|
genDisjunction vars spn
| Set.size spn == 1 =
case disjs of
[] -> Nothing
[x] -> Just x
_ -> error "CASL2TopSort.genDisjunction: this cannot happen"
| null disjs = Nothing
| otherwise = Just (disjunct disjs)
where disjs = foldl genConjunction [] (Set.toList spn)
genConjunction acc pns
| null conjs = acc
| otherwise = conjunct (reverse conjs) : acc
where conjs = foldl genPred [] (zip vars pns)
genPred acc (v, mpn) = maybe acc
(\ pn -> genPredication pn [v] : acc) mpn
{- | Each membership test of a subsort is transformed to a predication
of the corresponding unary predicate. Variables quantified over a
subsort yield a premise to the quantified formula that the
corresponding predicate holds. All typings are adjusted according
to the subsortmap and sort generation constraints are translated to
disjointness axioms. -}
| 963
| false
| true
| 0
| 12
| 288
| 258
| 124
| 134
| null | null |
amutake/imp
|
src/Main.hs
|
bsd-3-clause
|
help :: IO ()
help = putStr $ unlines
[ "Usage: imp [COMMAND or FILEPATH]"
, ""
, "Commands:"
, " help show this help text"
, " repl start REPL"
]
| 182
|
help :: IO ()
help = putStr $ unlines
[ "Usage: imp [COMMAND or FILEPATH]"
, ""
, "Commands:"
, " help show this help text"
, " repl start REPL"
]
| 182
|
help = putStr $ unlines
[ "Usage: imp [COMMAND or FILEPATH]"
, ""
, "Commands:"
, " help show this help text"
, " repl start REPL"
]
| 168
| false
| true
| 0
| 6
| 67
| 38
| 21
| 17
| null | null |
diku-kmc/repg
|
src/KMC/Kleenex/Parser.hs
|
mit
|
lexeme :: Parser a -> Parser a
lexeme p = p <* whiteSpace
| 57
|
lexeme :: Parser a -> Parser a
lexeme p = p <* whiteSpace
| 57
|
lexeme p = p <* whiteSpace
| 26
| false
| true
| 2
| 7
| 12
| 35
| 15
| 20
| null | null |
gridaphobe/ghc
|
compiler/hsSyn/HsExpr.hs
|
bsd-3-clause
|
matchContextErrString IfAlt = text "multi-way if"
| 64
|
matchContextErrString IfAlt = text "multi-way if"
| 64
|
matchContextErrString IfAlt = text "multi-way if"
| 64
| false
| false
| 0
| 5
| 20
| 12
| 5
| 7
| null | null |
deontologician/orbRPG
|
Game/OrbRPG/Combinations.hs
|
gpl-3.0
|
P Green @>> E Deuterium = L Lambda
| 34
|
P Green @>> E Deuterium = L Lambda
| 34
|
P Green @>> E Deuterium = L Lambda
| 34
| false
| false
| 0
| 6
| 7
| 21
| 8
| 13
| null | null |
sbditto85/parsedsltest
|
test/Generators.hs
|
apache-2.0
|
genJsonValueToParsed :: JsonValue -> JsonValue
genJsonValueToParsed (JsonValue strConcat) = JsonValue $ genStrConcatToParsed strConcat
| 134
|
genJsonValueToParsed :: JsonValue -> JsonValue
genJsonValueToParsed (JsonValue strConcat) = JsonValue $ genStrConcatToParsed strConcat
| 134
|
genJsonValueToParsed (JsonValue strConcat) = JsonValue $ genStrConcatToParsed strConcat
| 87
| false
| true
| 0
| 7
| 12
| 31
| 15
| 16
| null | null |
mdsteele/fallback
|
src/Fallback/State/Item.hs
|
gpl-3.0
|
weaponName NastyKnife = "Nasty Knife"
| 37
|
weaponName NastyKnife = "Nasty Knife"
| 37
|
weaponName NastyKnife = "Nasty Knife"
| 37
| false
| false
| 1
| 5
| 4
| 13
| 4
| 9
| null | null |
ComputationWithBoundedResources/ara-inference
|
doc/tpdb_trs/Haskell/basic_haskell/toEnum_3.hs
|
mit
|
primEqInt (Pos Zero) (Neg Zero) = MyTrue
| 40
|
primEqInt (Pos Zero) (Neg Zero) = MyTrue
| 40
|
primEqInt (Pos Zero) (Neg Zero) = MyTrue
| 40
| false
| false
| 0
| 7
| 6
| 23
| 11
| 12
| null | null |
shotofi/homepage
|
hs/homepage.hs
|
apache-2.0
|
fiTwoColumnLeft :: Context String -> Compiler (Item String)
fiTwoColumnLeft = applyTemplates "templates/two-column.html" fiTemplate
| 131
|
fiTwoColumnLeft :: Context String -> Compiler (Item String)
fiTwoColumnLeft = applyTemplates "templates/two-column.html" fiTemplate
| 131
|
fiTwoColumnLeft = applyTemplates "templates/two-column.html" fiTemplate
| 71
| false
| true
| 0
| 9
| 12
| 37
| 16
| 21
| null | null |
miniBill/entangle
|
src/exe/Examples.hs
|
mit
|
groverNaive :: (Qubit, Qubit, Qubit) -> Circ (Bit, Bit)
groverNaive (q1,q2,q3) = do
hadamard_at q1
hadamard_at q2
hadamard_at q3
--gate_X_at q2
qnot_at q3 `controlled` [q1, q2]
--gate_X_at q2
hadamard_at q1
hadamard_at q2
gate_X_at q1
gate_X_at q2
hadamard_at q2
qnot_at q2 `controlled` q1
hadamard_at q2
gate_X_at q1
gate_X_at q2
hadamard_at q1
hadamard_at q2
hadamard_at q3
measure (q1,q2)
| 467
|
groverNaive :: (Qubit, Qubit, Qubit) -> Circ (Bit, Bit)
groverNaive (q1,q2,q3) = do
hadamard_at q1
hadamard_at q2
hadamard_at q3
--gate_X_at q2
qnot_at q3 `controlled` [q1, q2]
--gate_X_at q2
hadamard_at q1
hadamard_at q2
gate_X_at q1
gate_X_at q2
hadamard_at q2
qnot_at q2 `controlled` q1
hadamard_at q2
gate_X_at q1
gate_X_at q2
hadamard_at q1
hadamard_at q2
hadamard_at q3
measure (q1,q2)
| 467
|
groverNaive (q1,q2,q3) = do
hadamard_at q1
hadamard_at q2
hadamard_at q3
--gate_X_at q2
qnot_at q3 `controlled` [q1, q2]
--gate_X_at q2
hadamard_at q1
hadamard_at q2
gate_X_at q1
gate_X_at q2
hadamard_at q2
qnot_at q2 `controlled` q1
hadamard_at q2
gate_X_at q1
gate_X_at q2
hadamard_at q1
hadamard_at q2
hadamard_at q3
measure (q1,q2)
| 411
| false
| true
| 0
| 8
| 131
| 173
| 78
| 95
| null | null |
AlexeyRaga/eta
|
compiler/ETA/TypeCheck/TcForeign.hs
|
bsd-3-clause
|
checkForeignRes :: Bool -> Bool -> (Type -> Validity) -> Type -> TcM ()
checkForeignRes nonIOResultOk checkSafe predResType ty
| Just (_, resType) <- tcSplitIOType_maybe ty
= do
traceTc "checkForeignRes[IO]" $ ppr resType
check (predResType resType) (illegalForeignTyErr result)
| Just (_, tagType, resType) <- tcSplitJavaType_maybe ty
= do
traceTc "checkForeignRes[Java]" $ ppr tagType <+> ppr resType
check (predResType resType) (illegalForeignTyErr result)
-- Case for non-IO result type with FFI Import
| not nonIOResultOk = addErrTc
. illegalForeignTyErr result
$ str "IO result type expected"
| otherwise = do
traceTc "checkForeignRes[Other]" $ ppr ty
dflags <- getDynFlags
case predResType ty of
-- Handle normal typecheck fail, we want to handle this first and
-- only report safe haskell errors if the normal type check is OK.
NotValid msg -> addErrTc $ illegalForeignTyErr result msg
-- handle safe infer fail
_ | checkSafe && safeInferOn dflags -> recordUnsafeInfer
-- handle safe language typecheck fail
_ | checkSafe && safeLanguageOn dflags ->
addErrTc $ illegalForeignTyErr result safeHsErr
-- sucess! non-IO return is fine
_ -> return ()
where safeHsErr = str $ "Safe Haskell is on, all FFI imports must be in the"
++ " IO monad"
| 1,454
|
checkForeignRes :: Bool -> Bool -> (Type -> Validity) -> Type -> TcM ()
checkForeignRes nonIOResultOk checkSafe predResType ty
| Just (_, resType) <- tcSplitIOType_maybe ty
= do
traceTc "checkForeignRes[IO]" $ ppr resType
check (predResType resType) (illegalForeignTyErr result)
| Just (_, tagType, resType) <- tcSplitJavaType_maybe ty
= do
traceTc "checkForeignRes[Java]" $ ppr tagType <+> ppr resType
check (predResType resType) (illegalForeignTyErr result)
-- Case for non-IO result type with FFI Import
| not nonIOResultOk = addErrTc
. illegalForeignTyErr result
$ str "IO result type expected"
| otherwise = do
traceTc "checkForeignRes[Other]" $ ppr ty
dflags <- getDynFlags
case predResType ty of
-- Handle normal typecheck fail, we want to handle this first and
-- only report safe haskell errors if the normal type check is OK.
NotValid msg -> addErrTc $ illegalForeignTyErr result msg
-- handle safe infer fail
_ | checkSafe && safeInferOn dflags -> recordUnsafeInfer
-- handle safe language typecheck fail
_ | checkSafe && safeLanguageOn dflags ->
addErrTc $ illegalForeignTyErr result safeHsErr
-- sucess! non-IO return is fine
_ -> return ()
where safeHsErr = str $ "Safe Haskell is on, all FFI imports must be in the"
++ " IO monad"
| 1,454
|
checkForeignRes nonIOResultOk checkSafe predResType ty
| Just (_, resType) <- tcSplitIOType_maybe ty
= do
traceTc "checkForeignRes[IO]" $ ppr resType
check (predResType resType) (illegalForeignTyErr result)
| Just (_, tagType, resType) <- tcSplitJavaType_maybe ty
= do
traceTc "checkForeignRes[Java]" $ ppr tagType <+> ppr resType
check (predResType resType) (illegalForeignTyErr result)
-- Case for non-IO result type with FFI Import
| not nonIOResultOk = addErrTc
. illegalForeignTyErr result
$ str "IO result type expected"
| otherwise = do
traceTc "checkForeignRes[Other]" $ ppr ty
dflags <- getDynFlags
case predResType ty of
-- Handle normal typecheck fail, we want to handle this first and
-- only report safe haskell errors if the normal type check is OK.
NotValid msg -> addErrTc $ illegalForeignTyErr result msg
-- handle safe infer fail
_ | checkSafe && safeInferOn dflags -> recordUnsafeInfer
-- handle safe language typecheck fail
_ | checkSafe && safeLanguageOn dflags ->
addErrTc $ illegalForeignTyErr result safeHsErr
-- sucess! non-IO return is fine
_ -> return ()
where safeHsErr = str $ "Safe Haskell is on, all FFI imports must be in the"
++ " IO monad"
| 1,382
| false
| true
| 2
| 14
| 402
| 335
| 154
| 181
| null | null |
keera-studios/hsQt
|
Qtc/Gui/QStyleOptionToolBar.hs
|
bsd-2-clause
|
setToolBarArea :: QStyleOptionToolBar a -> ((ToolBarArea)) -> IO ()
setToolBarArea x0 (x1)
= withObjectPtr x0 $ \cobj_x0 ->
qtc_QStyleOptionToolBar_setToolBarArea cobj_x0 (toCLong $ qEnum_toInt x1)
| 203
|
setToolBarArea :: QStyleOptionToolBar a -> ((ToolBarArea)) -> IO ()
setToolBarArea x0 (x1)
= withObjectPtr x0 $ \cobj_x0 ->
qtc_QStyleOptionToolBar_setToolBarArea cobj_x0 (toCLong $ qEnum_toInt x1)
| 203
|
setToolBarArea x0 (x1)
= withObjectPtr x0 $ \cobj_x0 ->
qtc_QStyleOptionToolBar_setToolBarArea cobj_x0 (toCLong $ qEnum_toInt x1)
| 135
| false
| true
| 0
| 10
| 29
| 68
| 34
| 34
| null | null |
lambdageek/insomnia
|
tests/TestToAST.hs
|
bsd-3-clause
|
units :: TestTree
units = testGroup "ToAST" [
infixParsingUnits
]
| 69
|
units :: TestTree
units = testGroup "ToAST" [
infixParsingUnits
]
| 69
|
units = testGroup "ToAST" [
infixParsingUnits
]
| 51
| false
| true
| 0
| 6
| 13
| 19
| 10
| 9
| null | null |
imeckler/mote
|
Mote.hs
|
bsd-3-clause
|
respond' :: Ref MoteState -> FromClient -> M ToClient
respond' stRef = \case
Load p -> do
loadFile stRef p
liftIO . forkIO $ do
fmap fileData (gReadRef stRef) >>= \case
Nothing -> return ()
Just (FileData {holesInfo=_}) ->
return ()
-- no idea if this is kosher
-- void . runGhc (Just libdir) $ runErrorT (mapM_ (getAndMemoizeSuggestions stRef) (M.elems holesInfo))
return Ok
NextHole (ClientState {path, cursorPos=(line,col)}) ->
getHoles stRef >>= \holes ->
let mh =
case dropWhile ((currPosLoc >=) . srcSpanStart) holes of
[] -> case holes of {[] -> Nothing; (h:_) -> Just h }
(h:_) -> Just h
in
maybe (throwError NoHole) (return . SetCursor . srcLocPos . srcSpanStart) mh
where
currPosLoc = mkSrcLoc (fsLit path) line col
-- inefficient
PrevHole (ClientState {path, cursorPos=(line, col)}) ->
getHoles stRef >>= \holes ->
let mxs = case takeWhile (< currPosSpan) holes of
[] -> case holes of {[] -> Nothing; _ -> Just holes}
xs -> Just xs
in
maybe (throwError NoHole) (return . SetCursor . srcLocPos . srcSpanStart . last) mxs
where
currPosSpan = srcLocSpan (mkSrcLoc (fsLit path) line col)
EnterHole (ClientState {..}) -> do
FileData {path=p} <- getFileDataErr stRef
when (p /= path) (loadFile stRef path)
mh <- getEnclosingHole stRef cursorPos
gModifyRef stRef (\st -> st { currentHole = mh })
return $ case mh of
Nothing -> SetInfoWindow "No Hole found"
Just _ -> Ok
GetHoleInfo (ClientState {..}) (HoleInfoOptions{..}) -> do
ahi@(AugmentedHoleInfo {holeInfo=hi}) <- getCurrentHoleErr stRef
fs <- lift getSessionDynFlags
let env = map (\(id,t) -> (occNameToString (getOccName id), showType fs t)) (holeEnv hi)
case sendOutputAsData of
True -> do
suggsJSON <-
if withSuggestions
then mkSuggsJSON <$> Mote.Suggest.getAndMemoizeSuggestions stRef ahi
else return []
return $
HoleInfoJSON . Aeson.object $
[ "environment" .= map (\(x, t) -> Aeson.object ["name" .= x, "type" .= t]) env
, "goal" .= Aeson.object ["name" .= holeNameString hi, "type" .= showType fs (holeType hi) ]
]
++ suggsJSON
where
mkSuggJSON (n, t) = Aeson.object ["name" .= occNameToString (occName n), "type" .= showType fs t]
mkSuggsJSON suggs = [ "suggestions" .= map mkSuggJSON suggs ]
False -> do
suggsStr <-
if withSuggestions
then mkSuggsStr <$> Mote.Suggest.getAndMemoizeSuggestions stRef ahi
else return ""
let goalStr = "Goal: " ++ holeNameString hi ++ " :: " ++ showType fs (holeType hi)
++ "\n" ++ replicate 40 '-'
envStr =
-- TODO: Wow, this is total for the strangest reason. If env
-- is empty then maxIdentLength never gets used to pad so
-- maximum doesn't fail.
let maxIdentLength = maximum $ map (\(x,_) -> length x) env in
unlines $ map (\(x, t) ->
take maxIdentLength (x ++ repeat ' ') ++ " :: " ++ t) env
return . SetInfoWindow $ unlines [goalStr, envStr, "", suggsStr]
where
mkSuggsStr suggs =
let heading = "Suggestions:\n" ++ replicate 40 '-' in
unlines (heading : map (\(n, t) -> (occNameToString . occName) n ++ " :: " ++ showType fs t) suggs)
Refine exprStr (ClientState {..}) -> do
hi <- getCurrentHoleErr stRef
expr' <- refine stRef exprStr
fs <- lift getSessionDynFlags
unqual <- lift getPrintUnqual
return $
Replace (toSpan . holeSpan $ holeInfo hi) path
(showSDocForUser fs unqual (ppr expr'))
SendStop -> return Stop
-- Precondition here: Hole has already been entered
CaseFurther var ClientState {} -> do
MoteState {..} <- gReadRef stRef
FileData {path, hsModule} <- getFileDataErr stRef
hi@(HoleInfo {holeEnv}) <- holeInfo <$> getCurrentHoleErr stRef
(id, ty) <- maybeThrow (NoVariable var) $
List.find (\(id,_) -> var == occNameToString (getOccName id)) holeEnv
expansions stRef (occNameToString (getOccName id)) ty (holeSpan hi) hsModule >>= \case
Nothing -> return (Error "Variable not found")
Just ((L sp _mg, mi), matches) -> do
fs <- lift getSessionDynFlags
unqual <- lift getPrintUnqual
let span = toSpan sp
indentLevel = subtract 1 . snd . fst $ span
indentTail [] = error "indentTail got []"
indentTail (s:ss) = s : map (replicate indentLevel ' ' ++) ss
showMatch :: HsMatchContext RdrName -> Match RdrName (LHsExpr RdrName) -> String
showMatch ctx = showSDocForUser fs unqual . pprMatch ctx
return $ case mi of
Equation (L _l name) ->
Replace (toSpan sp) path . unlines . indentTail $
map (showMatch (FunRhs name False)) matches
CaseBranch ->
-- TODO shouldn't always unlines. sometimes should be ; and {}
Replace (toSpan sp) path . unlines . indentTail $
map (showMatch CaseAlt) matches
SingleLambda _loc ->
Error "SingleLambda case expansion not yet implemented"
CaseOn exprStr (ClientState {..}) -> do
expr <- parseExpr exprStr
-- Should actually have general mechanism for getting the scope at
-- a point...
FileData {..} <- getFileDataErr stRef
ty <- getEnclosingHole stRef cursorPos >>= \case
Nothing -> hsExprType expr
Just hi -> inHoleEnv typecheckedModule (holeInfo hi) $ tcRnExprTc expr
let (line, col) = cursorPos
ms <- matchesForTypeAt stRef ty (mkSrcLoc (fsLit "") line col)
indentLevel <- liftIO $
LB8.length . LB8.takeWhile (== ' ') . (!! (line - 1)) . LB8.lines <$> LB8.readFile path
fs <- lift getSessionDynFlags
unqual <- lift getPrintUnqual
let indent n = (replicate n ' ' ++)
showMatch = showSDocForUser fs unqual . pprMatch (CaseAlt :: HsMatchContext RdrName)
return
. Insert cursorPos path
. unlines
$ ("case " ++ exprStr ++ " of") : map (indent (2 + fromIntegral indentLevel) . showMatch) ms
-- every message should really send current file name (ClientState) and
-- check if it matches the currently loaded file
GetType e -> do
fs <- lift getSessionDynFlags
x <- exprType e
unqual <- lift getPrintUnqual
return . SetInfoWindow . showSDocForUser fs unqual $ ppr x
Search {} -> return Ok -- TODO
| 6,719
|
respond' :: Ref MoteState -> FromClient -> M ToClient
respond' stRef = \case
Load p -> do
loadFile stRef p
liftIO . forkIO $ do
fmap fileData (gReadRef stRef) >>= \case
Nothing -> return ()
Just (FileData {holesInfo=_}) ->
return ()
-- no idea if this is kosher
-- void . runGhc (Just libdir) $ runErrorT (mapM_ (getAndMemoizeSuggestions stRef) (M.elems holesInfo))
return Ok
NextHole (ClientState {path, cursorPos=(line,col)}) ->
getHoles stRef >>= \holes ->
let mh =
case dropWhile ((currPosLoc >=) . srcSpanStart) holes of
[] -> case holes of {[] -> Nothing; (h:_) -> Just h }
(h:_) -> Just h
in
maybe (throwError NoHole) (return . SetCursor . srcLocPos . srcSpanStart) mh
where
currPosLoc = mkSrcLoc (fsLit path) line col
-- inefficient
PrevHole (ClientState {path, cursorPos=(line, col)}) ->
getHoles stRef >>= \holes ->
let mxs = case takeWhile (< currPosSpan) holes of
[] -> case holes of {[] -> Nothing; _ -> Just holes}
xs -> Just xs
in
maybe (throwError NoHole) (return . SetCursor . srcLocPos . srcSpanStart . last) mxs
where
currPosSpan = srcLocSpan (mkSrcLoc (fsLit path) line col)
EnterHole (ClientState {..}) -> do
FileData {path=p} <- getFileDataErr stRef
when (p /= path) (loadFile stRef path)
mh <- getEnclosingHole stRef cursorPos
gModifyRef stRef (\st -> st { currentHole = mh })
return $ case mh of
Nothing -> SetInfoWindow "No Hole found"
Just _ -> Ok
GetHoleInfo (ClientState {..}) (HoleInfoOptions{..}) -> do
ahi@(AugmentedHoleInfo {holeInfo=hi}) <- getCurrentHoleErr stRef
fs <- lift getSessionDynFlags
let env = map (\(id,t) -> (occNameToString (getOccName id), showType fs t)) (holeEnv hi)
case sendOutputAsData of
True -> do
suggsJSON <-
if withSuggestions
then mkSuggsJSON <$> Mote.Suggest.getAndMemoizeSuggestions stRef ahi
else return []
return $
HoleInfoJSON . Aeson.object $
[ "environment" .= map (\(x, t) -> Aeson.object ["name" .= x, "type" .= t]) env
, "goal" .= Aeson.object ["name" .= holeNameString hi, "type" .= showType fs (holeType hi) ]
]
++ suggsJSON
where
mkSuggJSON (n, t) = Aeson.object ["name" .= occNameToString (occName n), "type" .= showType fs t]
mkSuggsJSON suggs = [ "suggestions" .= map mkSuggJSON suggs ]
False -> do
suggsStr <-
if withSuggestions
then mkSuggsStr <$> Mote.Suggest.getAndMemoizeSuggestions stRef ahi
else return ""
let goalStr = "Goal: " ++ holeNameString hi ++ " :: " ++ showType fs (holeType hi)
++ "\n" ++ replicate 40 '-'
envStr =
-- TODO: Wow, this is total for the strangest reason. If env
-- is empty then maxIdentLength never gets used to pad so
-- maximum doesn't fail.
let maxIdentLength = maximum $ map (\(x,_) -> length x) env in
unlines $ map (\(x, t) ->
take maxIdentLength (x ++ repeat ' ') ++ " :: " ++ t) env
return . SetInfoWindow $ unlines [goalStr, envStr, "", suggsStr]
where
mkSuggsStr suggs =
let heading = "Suggestions:\n" ++ replicate 40 '-' in
unlines (heading : map (\(n, t) -> (occNameToString . occName) n ++ " :: " ++ showType fs t) suggs)
Refine exprStr (ClientState {..}) -> do
hi <- getCurrentHoleErr stRef
expr' <- refine stRef exprStr
fs <- lift getSessionDynFlags
unqual <- lift getPrintUnqual
return $
Replace (toSpan . holeSpan $ holeInfo hi) path
(showSDocForUser fs unqual (ppr expr'))
SendStop -> return Stop
-- Precondition here: Hole has already been entered
CaseFurther var ClientState {} -> do
MoteState {..} <- gReadRef stRef
FileData {path, hsModule} <- getFileDataErr stRef
hi@(HoleInfo {holeEnv}) <- holeInfo <$> getCurrentHoleErr stRef
(id, ty) <- maybeThrow (NoVariable var) $
List.find (\(id,_) -> var == occNameToString (getOccName id)) holeEnv
expansions stRef (occNameToString (getOccName id)) ty (holeSpan hi) hsModule >>= \case
Nothing -> return (Error "Variable not found")
Just ((L sp _mg, mi), matches) -> do
fs <- lift getSessionDynFlags
unqual <- lift getPrintUnqual
let span = toSpan sp
indentLevel = subtract 1 . snd . fst $ span
indentTail [] = error "indentTail got []"
indentTail (s:ss) = s : map (replicate indentLevel ' ' ++) ss
showMatch :: HsMatchContext RdrName -> Match RdrName (LHsExpr RdrName) -> String
showMatch ctx = showSDocForUser fs unqual . pprMatch ctx
return $ case mi of
Equation (L _l name) ->
Replace (toSpan sp) path . unlines . indentTail $
map (showMatch (FunRhs name False)) matches
CaseBranch ->
-- TODO shouldn't always unlines. sometimes should be ; and {}
Replace (toSpan sp) path . unlines . indentTail $
map (showMatch CaseAlt) matches
SingleLambda _loc ->
Error "SingleLambda case expansion not yet implemented"
CaseOn exprStr (ClientState {..}) -> do
expr <- parseExpr exprStr
-- Should actually have general mechanism for getting the scope at
-- a point...
FileData {..} <- getFileDataErr stRef
ty <- getEnclosingHole stRef cursorPos >>= \case
Nothing -> hsExprType expr
Just hi -> inHoleEnv typecheckedModule (holeInfo hi) $ tcRnExprTc expr
let (line, col) = cursorPos
ms <- matchesForTypeAt stRef ty (mkSrcLoc (fsLit "") line col)
indentLevel <- liftIO $
LB8.length . LB8.takeWhile (== ' ') . (!! (line - 1)) . LB8.lines <$> LB8.readFile path
fs <- lift getSessionDynFlags
unqual <- lift getPrintUnqual
let indent n = (replicate n ' ' ++)
showMatch = showSDocForUser fs unqual . pprMatch (CaseAlt :: HsMatchContext RdrName)
return
. Insert cursorPos path
. unlines
$ ("case " ++ exprStr ++ " of") : map (indent (2 + fromIntegral indentLevel) . showMatch) ms
-- every message should really send current file name (ClientState) and
-- check if it matches the currently loaded file
GetType e -> do
fs <- lift getSessionDynFlags
x <- exprType e
unqual <- lift getPrintUnqual
return . SetInfoWindow . showSDocForUser fs unqual $ ppr x
Search {} -> return Ok -- TODO
| 6,719
|
respond' stRef = \case
Load p -> do
loadFile stRef p
liftIO . forkIO $ do
fmap fileData (gReadRef stRef) >>= \case
Nothing -> return ()
Just (FileData {holesInfo=_}) ->
return ()
-- no idea if this is kosher
-- void . runGhc (Just libdir) $ runErrorT (mapM_ (getAndMemoizeSuggestions stRef) (M.elems holesInfo))
return Ok
NextHole (ClientState {path, cursorPos=(line,col)}) ->
getHoles stRef >>= \holes ->
let mh =
case dropWhile ((currPosLoc >=) . srcSpanStart) holes of
[] -> case holes of {[] -> Nothing; (h:_) -> Just h }
(h:_) -> Just h
in
maybe (throwError NoHole) (return . SetCursor . srcLocPos . srcSpanStart) mh
where
currPosLoc = mkSrcLoc (fsLit path) line col
-- inefficient
PrevHole (ClientState {path, cursorPos=(line, col)}) ->
getHoles stRef >>= \holes ->
let mxs = case takeWhile (< currPosSpan) holes of
[] -> case holes of {[] -> Nothing; _ -> Just holes}
xs -> Just xs
in
maybe (throwError NoHole) (return . SetCursor . srcLocPos . srcSpanStart . last) mxs
where
currPosSpan = srcLocSpan (mkSrcLoc (fsLit path) line col)
EnterHole (ClientState {..}) -> do
FileData {path=p} <- getFileDataErr stRef
when (p /= path) (loadFile stRef path)
mh <- getEnclosingHole stRef cursorPos
gModifyRef stRef (\st -> st { currentHole = mh })
return $ case mh of
Nothing -> SetInfoWindow "No Hole found"
Just _ -> Ok
GetHoleInfo (ClientState {..}) (HoleInfoOptions{..}) -> do
ahi@(AugmentedHoleInfo {holeInfo=hi}) <- getCurrentHoleErr stRef
fs <- lift getSessionDynFlags
let env = map (\(id,t) -> (occNameToString (getOccName id), showType fs t)) (holeEnv hi)
case sendOutputAsData of
True -> do
suggsJSON <-
if withSuggestions
then mkSuggsJSON <$> Mote.Suggest.getAndMemoizeSuggestions stRef ahi
else return []
return $
HoleInfoJSON . Aeson.object $
[ "environment" .= map (\(x, t) -> Aeson.object ["name" .= x, "type" .= t]) env
, "goal" .= Aeson.object ["name" .= holeNameString hi, "type" .= showType fs (holeType hi) ]
]
++ suggsJSON
where
mkSuggJSON (n, t) = Aeson.object ["name" .= occNameToString (occName n), "type" .= showType fs t]
mkSuggsJSON suggs = [ "suggestions" .= map mkSuggJSON suggs ]
False -> do
suggsStr <-
if withSuggestions
then mkSuggsStr <$> Mote.Suggest.getAndMemoizeSuggestions stRef ahi
else return ""
let goalStr = "Goal: " ++ holeNameString hi ++ " :: " ++ showType fs (holeType hi)
++ "\n" ++ replicate 40 '-'
envStr =
-- TODO: Wow, this is total for the strangest reason. If env
-- is empty then maxIdentLength never gets used to pad so
-- maximum doesn't fail.
let maxIdentLength = maximum $ map (\(x,_) -> length x) env in
unlines $ map (\(x, t) ->
take maxIdentLength (x ++ repeat ' ') ++ " :: " ++ t) env
return . SetInfoWindow $ unlines [goalStr, envStr, "", suggsStr]
where
mkSuggsStr suggs =
let heading = "Suggestions:\n" ++ replicate 40 '-' in
unlines (heading : map (\(n, t) -> (occNameToString . occName) n ++ " :: " ++ showType fs t) suggs)
Refine exprStr (ClientState {..}) -> do
hi <- getCurrentHoleErr stRef
expr' <- refine stRef exprStr
fs <- lift getSessionDynFlags
unqual <- lift getPrintUnqual
return $
Replace (toSpan . holeSpan $ holeInfo hi) path
(showSDocForUser fs unqual (ppr expr'))
SendStop -> return Stop
-- Precondition here: Hole has already been entered
CaseFurther var ClientState {} -> do
MoteState {..} <- gReadRef stRef
FileData {path, hsModule} <- getFileDataErr stRef
hi@(HoleInfo {holeEnv}) <- holeInfo <$> getCurrentHoleErr stRef
(id, ty) <- maybeThrow (NoVariable var) $
List.find (\(id,_) -> var == occNameToString (getOccName id)) holeEnv
expansions stRef (occNameToString (getOccName id)) ty (holeSpan hi) hsModule >>= \case
Nothing -> return (Error "Variable not found")
Just ((L sp _mg, mi), matches) -> do
fs <- lift getSessionDynFlags
unqual <- lift getPrintUnqual
let span = toSpan sp
indentLevel = subtract 1 . snd . fst $ span
indentTail [] = error "indentTail got []"
indentTail (s:ss) = s : map (replicate indentLevel ' ' ++) ss
showMatch :: HsMatchContext RdrName -> Match RdrName (LHsExpr RdrName) -> String
showMatch ctx = showSDocForUser fs unqual . pprMatch ctx
return $ case mi of
Equation (L _l name) ->
Replace (toSpan sp) path . unlines . indentTail $
map (showMatch (FunRhs name False)) matches
CaseBranch ->
-- TODO shouldn't always unlines. sometimes should be ; and {}
Replace (toSpan sp) path . unlines . indentTail $
map (showMatch CaseAlt) matches
SingleLambda _loc ->
Error "SingleLambda case expansion not yet implemented"
CaseOn exprStr (ClientState {..}) -> do
expr <- parseExpr exprStr
-- Should actually have general mechanism for getting the scope at
-- a point...
FileData {..} <- getFileDataErr stRef
ty <- getEnclosingHole stRef cursorPos >>= \case
Nothing -> hsExprType expr
Just hi -> inHoleEnv typecheckedModule (holeInfo hi) $ tcRnExprTc expr
let (line, col) = cursorPos
ms <- matchesForTypeAt stRef ty (mkSrcLoc (fsLit "") line col)
indentLevel <- liftIO $
LB8.length . LB8.takeWhile (== ' ') . (!! (line - 1)) . LB8.lines <$> LB8.readFile path
fs <- lift getSessionDynFlags
unqual <- lift getPrintUnqual
let indent n = (replicate n ' ' ++)
showMatch = showSDocForUser fs unqual . pprMatch (CaseAlt :: HsMatchContext RdrName)
return
. Insert cursorPos path
. unlines
$ ("case " ++ exprStr ++ " of") : map (indent (2 + fromIntegral indentLevel) . showMatch) ms
-- every message should really send current file name (ClientState) and
-- check if it matches the currently loaded file
GetType e -> do
fs <- lift getSessionDynFlags
x <- exprType e
unqual <- lift getPrintUnqual
return . SetInfoWindow . showSDocForUser fs unqual $ ppr x
Search {} -> return Ok -- TODO
| 6,665
| false
| true
| 0
| 29
| 1,997
| 2,219
| 1,089
| 1,130
| null | null |
vTurbine/ghc
|
compiler/hsSyn/HsPat.hs
|
bsd-3-clause
|
hsRecUpdFieldOcc :: HsRecField' (AmbiguousFieldOcc Id) arg -> LFieldOcc Id
hsRecUpdFieldOcc = fmap unambiguousFieldOcc . hsRecFieldLbl
| 134
|
hsRecUpdFieldOcc :: HsRecField' (AmbiguousFieldOcc Id) arg -> LFieldOcc Id
hsRecUpdFieldOcc = fmap unambiguousFieldOcc . hsRecFieldLbl
| 134
|
hsRecUpdFieldOcc = fmap unambiguousFieldOcc . hsRecFieldLbl
| 59
| false
| true
| 0
| 8
| 14
| 36
| 17
| 19
| null | null |
jmct/IterativeCompiler
|
frontend/Parser.hs
|
mit
|
second :: a -> b -> b
second a b = b
| 36
|
second :: a -> b -> b
second a b = b
| 36
|
second a b = b
| 14
| false
| true
| 0
| 8
| 11
| 30
| 13
| 17
| null | null |
motiz88/postgrest
|
test/Feature/StructureSpec.hs
|
mit
|
spec :: DbStructure -> H.Connection -> Spec
spec struct c = around (withApp cfgDefault struct c) $ do
describe "GET /" $ do
it "lists views in schema" $
request methodGet "/" [] ""
`shouldRespondWith` [json| [
{"schema":"test","name":"Escap3e;","insertable":true}
, {"schema":"test","name":"articleStars","insertable":true}
, {"schema":"test","name":"articles","insertable":true}
, {"schema":"test","name":"auto_incrementing_pk","insertable":true}
, {"schema":"test","name":"clients","insertable":true}
, {"schema":"test","name":"comments","insertable":true}
, {"schema":"test","name":"complex_items","insertable":true}
, {"schema":"test","name":"compound_pk","insertable":true}
, {"schema":"test","name":"ghostBusters","insertable":true}
, {"schema":"test","name":"has_count_column","insertable":false}
, {"schema":"test","name":"has_fk","insertable":true}
, {"schema":"test","name":"insertable_view_with_join","insertable":true}
, {"schema":"test","name":"insertonly","insertable":true}
, {"schema":"test","name":"items","insertable":true}
, {"schema":"test","name":"json","insertable":true}
, {"schema":"test","name":"materialized_view","insertable":false}
, {"schema":"test","name":"menagerie","insertable":true}
, {"schema":"test","name":"no_pk","insertable":true}
, {"schema":"test","name":"nullable_integer","insertable":true}
, {"schema":"test","name":"projects","insertable":true}
, {"schema":"test","name":"projects_view","insertable":true}
, {"schema":"test","name":"simple_pk","insertable":true}
, {"schema":"test","name":"tasks","insertable":true}
, {"schema":"test","name":"tsearch","insertable":true}
, {"schema":"test","name":"users","insertable":true}
, {"schema":"test","name":"users_projects","insertable":true}
, {"schema":"test","name":"users_tasks","insertable":true}
, {"schema":"test","name":"withUnique","insertable":true}
] |]
{matchStatus = 200}
it "lists only views user has permission to see" $ do
let auth = authHeaderJWT "eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJyb2xlIjoicG9zdGdyZXN0X3Rlc3RfYXV0aG9yIiwiaWQiOiJqZG9lIn0.y4vZuu1dDdwAl0-S00MCRWRYMlJ5YAMSir6Es6WtWx0"
request methodGet "/" [auth] ""
`shouldRespondWith` [json| [
{"schema":"test","name":"authors_only","insertable":true}
] |]
{matchStatus = 200}
describe "Table info" $ do
it "is available with OPTIONS verb" $
request methodOptions "/menagerie" [] "" `shouldRespondWith`
[json|
{
"pkey":["integer"],
"columns":[
{
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "integer",
"type": "integer",
"maxLen": null,
"enum": [],
"nullable": false,
"position": 1,
"references": null,
"default": null
}, {
"default": null,
"precision": 53,
"updatable": true,
"schema": "test",
"name": "double",
"type": "double precision",
"maxLen": null,
"enum": [],
"nullable": false,
"references": null,
"position": 2
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "varchar",
"type": "character varying",
"maxLen": null,
"enum": [],
"nullable": false,
"position": 3,
"references": null,
"default": null
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "boolean",
"type": "boolean",
"maxLen": null,
"enum": [],
"nullable": false,
"references": null,
"position": 4
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "date",
"type": "date",
"maxLen": null,
"enum": [],
"nullable": false,
"references": null,
"position": 5
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "money",
"type": "money",
"maxLen": null,
"enum": [],
"nullable": false,
"position": 6,
"references": null,
"default": null
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "enum",
"type": "USER-DEFINED",
"maxLen": null,
"enum": [
"foo",
"bar"
],
"nullable": false,
"position": 7,
"references": null,
"default": null
}
]
}
|]
it "it includes primary and foreign keys for views" $
request methodOptions "/projects_view" [] "" `shouldRespondWith`
[json|
{
"pkey":[
"id"
],
"columns":[
{
"references":null,
"default":null,
"precision":32,
"updatable":true,
"schema":"test",
"name":"id",
"type":"integer",
"maxLen":null,
"enum":[],
"nullable":true,
"position":1
},
{
"references":null,
"default":null,
"precision":null,
"updatable":true,
"schema":"test",
"name":"name",
"type":"text",
"maxLen":null,
"enum":[],
"nullable":true,
"position":2
},
{
"references": {
"schema":"test",
"column":"id",
"table":"clients"
},
"default":null,
"precision":32,
"updatable":true,
"schema":"test",
"name":"client_id",
"type":"integer",
"maxLen":null,
"enum":[],
"nullable":true,
"position":3
}
]
}
|]
it "includes foreign key data" $
request methodOptions "/has_fk" [] ""
`shouldRespondWith` [json|
{
"pkey": ["id"],
"columns":[
{
"default": "nextval('test.has_fk_id_seq'::regclass)",
"precision": 64,
"updatable": true,
"schema": "test",
"name": "id",
"type": "bigint",
"maxLen": null,
"nullable": false,
"position": 1,
"enum": [],
"references": null
}, {
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "auto_inc_fk",
"type": "integer",
"maxLen": null,
"nullable": true,
"position": 2,
"enum": [],
"references": {"schema":"test", "table": "auto_incrementing_pk", "column": "id"}
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "simple_fk",
"type": "character varying",
"maxLen": 255,
"nullable": true,
"position": 3,
"enum": [],
"references": {"schema":"test", "table": "simple_pk", "column": "k"}
}
]
}
|]
it "includes all information on views for renamed columns, and raises relations to correct schema" $
request methodOptions "/articleStars" [] ""
`shouldRespondWith` [json|
{
"pkey": [
"articleId",
"userId"
],
"columns": [
{
"references": {
"schema": "test",
"column": "id",
"table": "articles"
},
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "articleId",
"type": "integer",
"maxLen": null,
"enum": [],
"nullable": true,
"position": 1
},
{
"references": {
"schema": "test",
"column": "id",
"table": "users"
},
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "userId",
"type": "integer",
"maxLen": null,
"enum": [],
"nullable": true,
"position": 2
},
{
"references": null,
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "createdAt",
"type": "timestamp without time zone",
"maxLen": null,
"enum": [],
"nullable": true,
"position": 3
}
]
}
|]
it "errors for non existant tables" $
request methodOptions "/dne" [] "" `shouldRespondWith` 404
| 9,989
|
spec :: DbStructure -> H.Connection -> Spec
spec struct c = around (withApp cfgDefault struct c) $ do
describe "GET /" $ do
it "lists views in schema" $
request methodGet "/" [] ""
`shouldRespondWith` [json| [
{"schema":"test","name":"Escap3e;","insertable":true}
, {"schema":"test","name":"articleStars","insertable":true}
, {"schema":"test","name":"articles","insertable":true}
, {"schema":"test","name":"auto_incrementing_pk","insertable":true}
, {"schema":"test","name":"clients","insertable":true}
, {"schema":"test","name":"comments","insertable":true}
, {"schema":"test","name":"complex_items","insertable":true}
, {"schema":"test","name":"compound_pk","insertable":true}
, {"schema":"test","name":"ghostBusters","insertable":true}
, {"schema":"test","name":"has_count_column","insertable":false}
, {"schema":"test","name":"has_fk","insertable":true}
, {"schema":"test","name":"insertable_view_with_join","insertable":true}
, {"schema":"test","name":"insertonly","insertable":true}
, {"schema":"test","name":"items","insertable":true}
, {"schema":"test","name":"json","insertable":true}
, {"schema":"test","name":"materialized_view","insertable":false}
, {"schema":"test","name":"menagerie","insertable":true}
, {"schema":"test","name":"no_pk","insertable":true}
, {"schema":"test","name":"nullable_integer","insertable":true}
, {"schema":"test","name":"projects","insertable":true}
, {"schema":"test","name":"projects_view","insertable":true}
, {"schema":"test","name":"simple_pk","insertable":true}
, {"schema":"test","name":"tasks","insertable":true}
, {"schema":"test","name":"tsearch","insertable":true}
, {"schema":"test","name":"users","insertable":true}
, {"schema":"test","name":"users_projects","insertable":true}
, {"schema":"test","name":"users_tasks","insertable":true}
, {"schema":"test","name":"withUnique","insertable":true}
] |]
{matchStatus = 200}
it "lists only views user has permission to see" $ do
let auth = authHeaderJWT "eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJyb2xlIjoicG9zdGdyZXN0X3Rlc3RfYXV0aG9yIiwiaWQiOiJqZG9lIn0.y4vZuu1dDdwAl0-S00MCRWRYMlJ5YAMSir6Es6WtWx0"
request methodGet "/" [auth] ""
`shouldRespondWith` [json| [
{"schema":"test","name":"authors_only","insertable":true}
] |]
{matchStatus = 200}
describe "Table info" $ do
it "is available with OPTIONS verb" $
request methodOptions "/menagerie" [] "" `shouldRespondWith`
[json|
{
"pkey":["integer"],
"columns":[
{
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "integer",
"type": "integer",
"maxLen": null,
"enum": [],
"nullable": false,
"position": 1,
"references": null,
"default": null
}, {
"default": null,
"precision": 53,
"updatable": true,
"schema": "test",
"name": "double",
"type": "double precision",
"maxLen": null,
"enum": [],
"nullable": false,
"references": null,
"position": 2
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "varchar",
"type": "character varying",
"maxLen": null,
"enum": [],
"nullable": false,
"position": 3,
"references": null,
"default": null
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "boolean",
"type": "boolean",
"maxLen": null,
"enum": [],
"nullable": false,
"references": null,
"position": 4
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "date",
"type": "date",
"maxLen": null,
"enum": [],
"nullable": false,
"references": null,
"position": 5
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "money",
"type": "money",
"maxLen": null,
"enum": [],
"nullable": false,
"position": 6,
"references": null,
"default": null
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "enum",
"type": "USER-DEFINED",
"maxLen": null,
"enum": [
"foo",
"bar"
],
"nullable": false,
"position": 7,
"references": null,
"default": null
}
]
}
|]
it "it includes primary and foreign keys for views" $
request methodOptions "/projects_view" [] "" `shouldRespondWith`
[json|
{
"pkey":[
"id"
],
"columns":[
{
"references":null,
"default":null,
"precision":32,
"updatable":true,
"schema":"test",
"name":"id",
"type":"integer",
"maxLen":null,
"enum":[],
"nullable":true,
"position":1
},
{
"references":null,
"default":null,
"precision":null,
"updatable":true,
"schema":"test",
"name":"name",
"type":"text",
"maxLen":null,
"enum":[],
"nullable":true,
"position":2
},
{
"references": {
"schema":"test",
"column":"id",
"table":"clients"
},
"default":null,
"precision":32,
"updatable":true,
"schema":"test",
"name":"client_id",
"type":"integer",
"maxLen":null,
"enum":[],
"nullable":true,
"position":3
}
]
}
|]
it "includes foreign key data" $
request methodOptions "/has_fk" [] ""
`shouldRespondWith` [json|
{
"pkey": ["id"],
"columns":[
{
"default": "nextval('test.has_fk_id_seq'::regclass)",
"precision": 64,
"updatable": true,
"schema": "test",
"name": "id",
"type": "bigint",
"maxLen": null,
"nullable": false,
"position": 1,
"enum": [],
"references": null
}, {
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "auto_inc_fk",
"type": "integer",
"maxLen": null,
"nullable": true,
"position": 2,
"enum": [],
"references": {"schema":"test", "table": "auto_incrementing_pk", "column": "id"}
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "simple_fk",
"type": "character varying",
"maxLen": 255,
"nullable": true,
"position": 3,
"enum": [],
"references": {"schema":"test", "table": "simple_pk", "column": "k"}
}
]
}
|]
it "includes all information on views for renamed columns, and raises relations to correct schema" $
request methodOptions "/articleStars" [] ""
`shouldRespondWith` [json|
{
"pkey": [
"articleId",
"userId"
],
"columns": [
{
"references": {
"schema": "test",
"column": "id",
"table": "articles"
},
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "articleId",
"type": "integer",
"maxLen": null,
"enum": [],
"nullable": true,
"position": 1
},
{
"references": {
"schema": "test",
"column": "id",
"table": "users"
},
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "userId",
"type": "integer",
"maxLen": null,
"enum": [],
"nullable": true,
"position": 2
},
{
"references": null,
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "createdAt",
"type": "timestamp without time zone",
"maxLen": null,
"enum": [],
"nullable": true,
"position": 3
}
]
}
|]
it "errors for non existant tables" $
request methodOptions "/dne" [] "" `shouldRespondWith` 404
| 9,989
|
spec struct c = around (withApp cfgDefault struct c) $ do
describe "GET /" $ do
it "lists views in schema" $
request methodGet "/" [] ""
`shouldRespondWith` [json| [
{"schema":"test","name":"Escap3e;","insertable":true}
, {"schema":"test","name":"articleStars","insertable":true}
, {"schema":"test","name":"articles","insertable":true}
, {"schema":"test","name":"auto_incrementing_pk","insertable":true}
, {"schema":"test","name":"clients","insertable":true}
, {"schema":"test","name":"comments","insertable":true}
, {"schema":"test","name":"complex_items","insertable":true}
, {"schema":"test","name":"compound_pk","insertable":true}
, {"schema":"test","name":"ghostBusters","insertable":true}
, {"schema":"test","name":"has_count_column","insertable":false}
, {"schema":"test","name":"has_fk","insertable":true}
, {"schema":"test","name":"insertable_view_with_join","insertable":true}
, {"schema":"test","name":"insertonly","insertable":true}
, {"schema":"test","name":"items","insertable":true}
, {"schema":"test","name":"json","insertable":true}
, {"schema":"test","name":"materialized_view","insertable":false}
, {"schema":"test","name":"menagerie","insertable":true}
, {"schema":"test","name":"no_pk","insertable":true}
, {"schema":"test","name":"nullable_integer","insertable":true}
, {"schema":"test","name":"projects","insertable":true}
, {"schema":"test","name":"projects_view","insertable":true}
, {"schema":"test","name":"simple_pk","insertable":true}
, {"schema":"test","name":"tasks","insertable":true}
, {"schema":"test","name":"tsearch","insertable":true}
, {"schema":"test","name":"users","insertable":true}
, {"schema":"test","name":"users_projects","insertable":true}
, {"schema":"test","name":"users_tasks","insertable":true}
, {"schema":"test","name":"withUnique","insertable":true}
] |]
{matchStatus = 200}
it "lists only views user has permission to see" $ do
let auth = authHeaderJWT "eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJyb2xlIjoicG9zdGdyZXN0X3Rlc3RfYXV0aG9yIiwiaWQiOiJqZG9lIn0.y4vZuu1dDdwAl0-S00MCRWRYMlJ5YAMSir6Es6WtWx0"
request methodGet "/" [auth] ""
`shouldRespondWith` [json| [
{"schema":"test","name":"authors_only","insertable":true}
] |]
{matchStatus = 200}
describe "Table info" $ do
it "is available with OPTIONS verb" $
request methodOptions "/menagerie" [] "" `shouldRespondWith`
[json|
{
"pkey":["integer"],
"columns":[
{
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "integer",
"type": "integer",
"maxLen": null,
"enum": [],
"nullable": false,
"position": 1,
"references": null,
"default": null
}, {
"default": null,
"precision": 53,
"updatable": true,
"schema": "test",
"name": "double",
"type": "double precision",
"maxLen": null,
"enum": [],
"nullable": false,
"references": null,
"position": 2
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "varchar",
"type": "character varying",
"maxLen": null,
"enum": [],
"nullable": false,
"position": 3,
"references": null,
"default": null
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "boolean",
"type": "boolean",
"maxLen": null,
"enum": [],
"nullable": false,
"references": null,
"position": 4
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "date",
"type": "date",
"maxLen": null,
"enum": [],
"nullable": false,
"references": null,
"position": 5
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "money",
"type": "money",
"maxLen": null,
"enum": [],
"nullable": false,
"position": 6,
"references": null,
"default": null
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "enum",
"type": "USER-DEFINED",
"maxLen": null,
"enum": [
"foo",
"bar"
],
"nullable": false,
"position": 7,
"references": null,
"default": null
}
]
}
|]
it "it includes primary and foreign keys for views" $
request methodOptions "/projects_view" [] "" `shouldRespondWith`
[json|
{
"pkey":[
"id"
],
"columns":[
{
"references":null,
"default":null,
"precision":32,
"updatable":true,
"schema":"test",
"name":"id",
"type":"integer",
"maxLen":null,
"enum":[],
"nullable":true,
"position":1
},
{
"references":null,
"default":null,
"precision":null,
"updatable":true,
"schema":"test",
"name":"name",
"type":"text",
"maxLen":null,
"enum":[],
"nullable":true,
"position":2
},
{
"references": {
"schema":"test",
"column":"id",
"table":"clients"
},
"default":null,
"precision":32,
"updatable":true,
"schema":"test",
"name":"client_id",
"type":"integer",
"maxLen":null,
"enum":[],
"nullable":true,
"position":3
}
]
}
|]
it "includes foreign key data" $
request methodOptions "/has_fk" [] ""
`shouldRespondWith` [json|
{
"pkey": ["id"],
"columns":[
{
"default": "nextval('test.has_fk_id_seq'::regclass)",
"precision": 64,
"updatable": true,
"schema": "test",
"name": "id",
"type": "bigint",
"maxLen": null,
"nullable": false,
"position": 1,
"enum": [],
"references": null
}, {
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "auto_inc_fk",
"type": "integer",
"maxLen": null,
"nullable": true,
"position": 2,
"enum": [],
"references": {"schema":"test", "table": "auto_incrementing_pk", "column": "id"}
}, {
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "simple_fk",
"type": "character varying",
"maxLen": 255,
"nullable": true,
"position": 3,
"enum": [],
"references": {"schema":"test", "table": "simple_pk", "column": "k"}
}
]
}
|]
it "includes all information on views for renamed columns, and raises relations to correct schema" $
request methodOptions "/articleStars" [] ""
`shouldRespondWith` [json|
{
"pkey": [
"articleId",
"userId"
],
"columns": [
{
"references": {
"schema": "test",
"column": "id",
"table": "articles"
},
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "articleId",
"type": "integer",
"maxLen": null,
"enum": [],
"nullable": true,
"position": 1
},
{
"references": {
"schema": "test",
"column": "id",
"table": "users"
},
"default": null,
"precision": 32,
"updatable": true,
"schema": "test",
"name": "userId",
"type": "integer",
"maxLen": null,
"enum": [],
"nullable": true,
"position": 2
},
{
"references": null,
"default": null,
"precision": null,
"updatable": true,
"schema": "test",
"name": "createdAt",
"type": "timestamp without time zone",
"maxLen": null,
"enum": [],
"nullable": true,
"position": 3
}
]
}
|]
it "errors for non existant tables" $
request methodOptions "/dne" [] "" `shouldRespondWith` 404
| 9,945
| false
| true
| 0
| 17
| 4,083
| 314
| 162
| 152
| null | null |
fmapfmapfmap/amazonka
|
amazonka-iam/gen/Network/AWS/IAM/Types/Product.hs
|
mpl-2.0
|
-- | The contents of the public key certificate chain.
sCertificateChain :: Lens' ServerCertificate (Maybe Text)
sCertificateChain = lens _sCertificateChain (\ s a -> s{_sCertificateChain = a})
| 193
|
sCertificateChain :: Lens' ServerCertificate (Maybe Text)
sCertificateChain = lens _sCertificateChain (\ s a -> s{_sCertificateChain = a})
| 138
|
sCertificateChain = lens _sCertificateChain (\ s a -> s{_sCertificateChain = a})
| 80
| true
| true
| 0
| 9
| 26
| 46
| 25
| 21
| null | null |
pgreze/7languages7weeks
|
haskell/day2.hs
|
mit
|
lstripWith chars = dropWhile (\x -> any (== x) chars)
| 53
|
lstripWith chars = dropWhile (\x -> any (== x) chars)
| 53
|
lstripWith chars = dropWhile (\x -> any (== x) chars)
| 53
| false
| false
| 0
| 9
| 9
| 29
| 15
| 14
| null | null |
slasser/AllStar
|
ParserGenerator/AtnToDot.hs
|
bsd-3-clause
|
-- Took these from an online example
createImage :: PrintDotRepr dg n => FilePath -> dg n -> IO FilePath
createImage n g = createImageInDir "." n Png g
| 151
|
createImage :: PrintDotRepr dg n => FilePath -> dg n -> IO FilePath
createImage n g = createImageInDir "." n Png g
| 114
|
createImage n g = createImageInDir "." n Png g
| 46
| true
| true
| 0
| 8
| 28
| 49
| 23
| 26
| null | null |
brendanhay/gogol
|
gogol-books/gen/Network/Google/Books/Types/Product.hs
|
mpl-2.0
|
-- | Creates a value of 'VolumeVolumeInfoIndustryIdentifiersItem' with the minimum fields required to make a request.
--
-- Use one of the following lenses to modify other fields as desired:
--
-- * 'vviiiiIdentifier'
--
-- * 'vviiiiType'
volumeVolumeInfoIndustryIdentifiersItem
:: VolumeVolumeInfoIndustryIdentifiersItem
volumeVolumeInfoIndustryIdentifiersItem =
VolumeVolumeInfoIndustryIdentifiersItem'
{_vviiiiIdentifier = Nothing, _vviiiiType = Nothing}
| 467
|
volumeVolumeInfoIndustryIdentifiersItem
:: VolumeVolumeInfoIndustryIdentifiersItem
volumeVolumeInfoIndustryIdentifiersItem =
VolumeVolumeInfoIndustryIdentifiersItem'
{_vviiiiIdentifier = Nothing, _vviiiiType = Nothing}
| 228
|
volumeVolumeInfoIndustryIdentifiersItem =
VolumeVolumeInfoIndustryIdentifiersItem'
{_vviiiiIdentifier = Nothing, _vviiiiType = Nothing}
| 141
| true
| true
| 1
| 7
| 59
| 39
| 23
| 16
| null | null |
davidfontenot/haskell-hashtag-viewer
|
src/WebServer.hs
|
mit
|
getPortFromArgs :: (Num a) => Map.Map String String -> a
getPortFromArgs argMap = case Map.lookup "port" argMap of
Just port -> fromIntegral (read port :: Integer)
Nothing -> 8001
| 233
|
getPortFromArgs :: (Num a) => Map.Map String String -> a
getPortFromArgs argMap = case Map.lookup "port" argMap of
Just port -> fromIntegral (read port :: Integer)
Nothing -> 8001
| 233
|
getPortFromArgs argMap = case Map.lookup "port" argMap of
Just port -> fromIntegral (read port :: Integer)
Nothing -> 8001
| 176
| false
| true
| 0
| 10
| 82
| 72
| 35
| 37
| null | null |
ghcjs/ghcjs
|
src/Gen2/Optimizer.hs
|
mit
|
intInfixOp GeOp i1 i2 = eBool (i1 >= i2)
| 49
|
intInfixOp GeOp i1 i2 = eBool (i1 >= i2)
| 49
|
intInfixOp GeOp i1 i2 = eBool (i1 >= i2)
| 49
| false
| false
| 0
| 7
| 17
| 23
| 11
| 12
| null | null |
knotman90/haskell-tutorial
|
code/Problem55/Problem55.hs
|
gpl-2.0
|
--a smarter solution involves the fact that a number that is NOT liry produces a chain of number that are not liry themself and should not be tested! This approach is Not needed here for 10000 numbers. Brute force works in milliseconds.
--Brute Force solution
solve :: Int
solve = let vals = (map (isLiry) [1..9999]) in length $ filter (==True) vals
| 350
|
solve :: Int
solve = let vals = (map (isLiry) [1..9999]) in length $ filter (==True) vals
| 89
|
solve = let vals = (map (isLiry) [1..9999]) in length $ filter (==True) vals
| 76
| true
| true
| 0
| 11
| 62
| 59
| 30
| 29
| null | null |
nickspinale/aether
|
src/Network/Aether/RW.hs
|
mit
|
buildWant :: Want -> B.ByteString
buildWant N4 = "n4"
| 53
|
buildWant :: Want -> B.ByteString
buildWant N4 = "n4"
| 53
|
buildWant N4 = "n4"
| 19
| false
| true
| 0
| 8
| 8
| 26
| 11
| 15
| null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.