magic.data.common 16.1.9

There is a newer version of this package available.
See the version list below for details.
dotnet add package magic.data.common --version 16.1.9                
NuGet\Install-Package magic.data.common -Version 16.1.9                
This command is intended to be used within the Package Manager Console in Visual Studio, as it uses the NuGet module's version of Install-Package.
<PackageReference Include="magic.data.common" Version="16.1.9" />                
For projects that support PackageReference, copy this XML node into the project file to reference the package.
paket add magic.data.common --version 16.1.9                
#r "nuget: magic.data.common, 16.1.9"                
#r directive can be used in F# Interactive and Polyglot Notebooks. Copy this into the interactive tool or source code of the script to reference the package.
// Install magic.data.common as a Cake Addin
#addin nuget:?package=magic.data.common&version=16.1.9

// Install magic.data.common as a Cake Tool
#tool nuget:?package=magic.data.common&version=16.1.9                

magic.data.common - Hyperlambda slots to access your SQL database

The magic.data.commonproject is the generic data adapter, that transform dynamically from a lambda node structure into SQL, and polymorphistically invokes your specialised data adapter, resulting in SQL statements executed towards your database type of choice. In addition, it contains helper slots to give you more "raw" database access, in addition to slots helping you to open database connections, create transactions, execute SQL, etc.

How to use [sql.*]

These slots never executes SQL towards your data adapter, but rather simply generates your SQL, and returns the results of the SQL generation process back to you. They're mostly intended for debugging purposes, and/or learning purposes, and can be interchanged with their [data.xxx] equivalent, and/or their [mysql.xxx]/[mssql.xxx]/[pgsql.xxx] equivalent, etc. You can substitute these slots with for instance [data.xxx] if you wish to actually execute some SQL towards your database adapter of choice.

How to use [data.*]

All of the [data.xxx] slots are actually just polymorphistically evaluating your specialised adapter's slots, such as for instance [data.connect], that will invoke [mysql.connect] if this is your default database. However, all of these lots can also be given an explicit [database-type] argument, being for instance "mssql", allowing you to choose a database type explicitly as you invoke it.

If you don't provide an explicit [database-type] argument to these slots, the default database type will be retrieved from your "appsettings.json" file, from the magic:databases:default value, and substitute the "data" parts with the value found from your configuration setting, to invoke your specialised implementation. For instance, if you're using MySQL as your default database type, and you invoke for instance [data.connect], this will transform into an invocation to [mysql.connect], allowing you to use generic database slot invocations, ignoring your database type, creating the correct SQL dialect for you automagically.

How to use [data.connect]

This slot will open a database connection for you. You can pass in a complete connection string (not recommended), or only the database name if you wish. If you pass in only the database name, the generic connection string for your database type of choice from your "appsettings.json" file will be used, substituting its {database} parts with the database of your choice. Inside of this slot, which actually is a lambda [eval] invocation, you can use any of the other slots, requiring an existing and open database connection to function. You can see an example below.

data.connect:sakila
   data.read
      table:actor

Since the [data.connect] slot actually takes a lambda object, you can also add any amount of other lambda invocations inside of the lambda object supplied to the slot, allowing you to for instance create loops, conditional executions, etc, inside of your invocation to [data.connect]. This is also true for all other slots taking a lambda object, such as for instance [data.transaction.create]. Inside your lambda object, an invocation towards your database such as e.g. [data.read], will be using this database connection, as long as the type of database is matching. The database connection will be kept open, and implicitly used, for the entirety of the lambda object. If you need another database connection inside of your lambda object, you'll need to nest [data.connect] invocations. You can also explicitly choose which connection string to use as you open a connection, by separating the connection string and the database name by a | symbol, and wrapping your entire value inside of brackets. If you have a connection string in your "appsettings.json" file named for instance "foo", and this connection string points to a server instance having a database named "bar", you could open a connection to this database using something resembling the following.

data.connect:[foo|bar]

This is why your connection strings should contain the {database} as a generic argument, since the slot substitutes the {database} parts dynamically as you create new connections. Below is an example of such a connection string configuration setting.

"generic": "Server=localhost\\SQLEXPRESS;Database={database};Trusted_Connection=True;"

How to use [data.select]

This slot allows you to pass in any arbitrary SQL you wish, and evaluate it to a DataReader, and return all records as a lambda object. You can find an example below.

data.connect:sakila
   data.select:select first_name, last_name from actor limit 5

Assuming you have the "sakila" database from Oracle installed in your database, and your default database type is MySQL, the result of the above will end up looking like the following.

data.connect
   data.select
      ""
         first_name:PENELOPE
         last_name:GUINESS
      ""
         first_name:NICK
         last_name:WAHLBERG
      ""
         first_name:ED
         last_name:CHASE
      ""
         first_name:JENNIFER
         last_name:DAVIS
      ""
         first_name:JOHNNY
         last_name:LOLLOBRIGIDA

This slot requires SQL resembling your specialised database type of dialect, and will not in any ways transpile the SQL towards your specific underlaying database type's SQL dialect. If you can, you should rather use [data.read], to avoid lockin towards a specific database vendor's SQL dialect. You can also select multiple result sets if you have batch type of SQL statements, containing multiple SQL statements, and you want to return the result of all SQL statements you're executing. You do this by providing a [multiple-result-sets] argument and set its value to boolean true. If you do this, the slot will return an array of arrays, one outer array for each result set your SQL generates. Below is an example.

data.select:"select * from table1; select * from table2;"
   multiple-result-sets:bool:true

Which would result in something resembling the following.

   data.select
      ""
         ""
            field1:foo1
            field2:bar1
         ""
            field1:foo2
            field2:bar2
      ""
         ""
            field1:foo1
            field2:bar1
         ""
            field1:foo2
            field2:bar2

How to use [data.scalar]

This slot is similar to the [data.select] slot, but will only return one value as the value of its node after execution. This slot is typically used for aggregate results. You can see an example below.

data.connect:sakila
   data.scalar:select count(*) from actor

After execution, your result will resemble the following.

data.connect
   data.scalar:long:200

Yet again you should prefer the [data.*] slots if you can.

How to use [data.execute]

This slot should be used if you don't expect any type of result at all, such as in for instance delete or update invocations, where you don't care about the result of the operation. You can find an example below.

data.connect:sakila

   // Notice, will throw! (hopefully!)
   data.execute:delete from non_existing_table

Yet again, prefer [data.delete] if you can.

Database transactions

Although you should be careful with database transactions, sometimes you really need them. For those cases you can use the following 3 slots to create, rollback, and/or commit transactions towards your underlaying database.

  • [data.transaction.create] - Creates a new database transaction
  • [data.transaction.commit] - Commits an existing open transaction
  • [data.transaction.rollback] - Roll back an existing open transaction

Notice - The default logic for a database transaction, is that unless it's explicitly committed before leaving scope, it will roll back by default. Below is an example of a transaction that will rollback, since it's not explicitly commited before leaving scope.

data.connect:sakila
   data.transaction.create
      data.execute:delete from film_actor

      /*
       * If you uncomment the line below the
       * transaction will be committed, resulting
       * in that everything from your film_actor
       * table will be deleted.
       */
      //data.transaction.commit

data.connect:sakila

   /*
    * Notice, this still returns 5462 items, since
    * transaction was implicitly rolled back above.
    */
   data.scalar:select count(*) from film_actor

A transaction typically follows your connection, implying to count items after the transaction has been rolled back, we'll need a new connection, as the above example illustrates.

How to use [sql.*] slots

All of these slots simply generates SQL for you, using the generic SQL dialect syntax, which might or might not work for your database adapter of choice. This allows you to create SQL statements without executing anything towards your database. This allows you to play around with the syntax, to understand how it works, and see how some semantic graph object results in an SQL statement before using it. All of these slots have [data.*] equivalent slots, which again polymorphistically invokes your specialised data adapter's equivalent, and/or can be parametrised with a database type - Which again resolves to the [mysql.*] equivalent if you supply "mysql" as your [database-type], and/or MySQL is your default database type as configured in your "appsettings.json" file. Hence, the documentation for these slots is also the documentation for your [data.*] slots.

How to use [sql.create]

This slot will generate the SQL necessary to insert a record into a database for you. Besides the table argument, this slot can only be given one argument, which is [values]. Below is an example of usage.

sql.create
   table:table1
   values
      field1:howdy
      field2:world

Notice, to avoid SQL injection attacks, this slot will always return parameters expected to be passed in from any potentially malicious clients as SQL parameters - Hence, the complete returned value of the above Hyperlambda will be as follows.

sql.create:insert into 'table1' ('field1', 'field2') values (@0, @1)
   @0:howdy
   @1:world

The basic idea is that everything that might be dynamically injected into your data access layer, should be consumed as SqlParameters, or something equivalent, to prevent SQL injection attacks towards your database. This is true for all arguments passed in as data for all slots in the project. The slot will in its specialized implementations return the ID of the inserted record if possible, unless you explicitly parametrize it with a [return-id] argument and set its value to boolean false.

How to use [sql.read]

This slot requires only one mandatory argument, being your table name. The slot creates a select SQL statement for you. An example can be found below.

sql.read
   table:foo

The above will result in the following SQL returned to you. If you're using the special implementations, such as e.g. [data.read], and/or [mssql.read] - The returned SQL might vary according to your dialect. But the results of executing the SQL will be the same.

select * from 'foo' limit 25

To avoid accidentally exhausting your database, this slot will by default limit your result set to 25 records unless explicitly overridden by your code with a [limit] argument. You can optionally supply the following arguments to this slot.

  • [columns] - Columns to select
  • [order] - Which column(s) to order the results by. You can supply multiple [order] arguments
  • [direction] - Which direction to order your columns
  • [limit] - How many records to return, default is 25. Set this value to -1 to avoid having the parser inject its default value
  • [offset] - Offset of where to start returning records
  • [where] - Where condition
  • [join] - Join condition
  • [group] - Group by declaration

For instance, to select only the "field1" column and the "field2" column from "table1", and ordering descendingly by "field3" - You can use something resembling the following.

sql.read
   table:table1
   columns
      field1
      field2
   order:field3
   direction:desc

This will result in the following SQL returned.

select 'field1','field2' from 'table1' order by 'field3' desc limit 25

The [direction] argument can only be either "asc" or "desc", implying ascending or descending. You can also supply multiple ordering columns, by separating them by ",". See an example below, which also specifies what table to use while ordering your results.

sql.read
   table:table1
   order:table1.field1, table1.field2

The above will result in the following SQL.

select * from 'table1' order by 'table1'.'field1' asc,'table1'.'field2' asc limit 25

And finally you can provide the [direction] argument on separate [order] arguments, such as the following illustrates.

sql.read
   table:table1

   order:field1
      direction:asc

   order:field2
      direction:desc

If you do, the ordering is sequentially applied.

Aggregate results

You can also create aggregate results, by simply adding your aggregate as your column, such as the following illustrates. The reasons why this works, is because if the SQL generator finds a paranthesis in your column declaration, it will simply ignore parsing that column altogether, and directly inject it into the resulting SQL's columns declaration.

sql.read
   table:table1
   limit:-1
   columns
      count(*)
        as:count

The above will result in the following SQL.

select count(*) as count from 'table1'

By setting [limit] to "-1", like we do above, we avoid adding the limit parts to our SQL. Unless you explicitly specify a limit, the default value will always be 25 to avoid accidentally exhausting your database, and/or web server, by selecting all records from a table with millions of records.

Paging

To page your [sql.read] results, use [limit] and [offset], such as the following illustrates. Even though we use "limit" and "offset", the correct syntax will be applied for your database type, depending upon which database type you're using - Implying for Microsoft SQL Server, it will inject MS SQL dialect, and not MySQL dialect. But the syntax for your lambda object still remains the same, making it simpler to create SQL dialect valid for your specific database type 100% transparently.

sql.read
   table:table1
   offset:5
   limit:10

The above will return the following SQL select * from 'table1' limit 10 offset 5. If you run the above lambda towards Microsoft SQL server, SQL syntax specific for MS SQL will be generated.

Aliasing column results

You can also extract columns with an alias, "renaming" the column in its result, such as the following illustrates.

sql.read
   table:table1
   columns

      table1.foo1
         as:howdy

      table1.foo2
         as:world

The above Hyperlambda will result in the following SQL.

select 'table1'.'foo1' as 'howdy','table1'.'foo2' as 'world'
   from 'table1' limit 25

Effectively resulting in that you'll have two columns returned after executing the above SQL, which are howdy and world. Combining this with the join features from this project, allows you to create any type of "projections" you wish.

Joins

The project supports joins by parametrizing your [sql.read] invocation with [join] arguments, beneath your [table] argument(s). You can only add [join] beneath [table] for [sql.read] invocations though. If you have created the Sakila example database from Oracle, and you're using MySQL as your default database type, you can execute the following MySQL join SQL statement to see a fairly complex recursive join.

data.connect:sakila
   data.read

      columns
         title
         description
         last_name
         first_name

      table:film

         join:film_actor
            type:inner
            on
               and
                  film.film_id:film_actor.film_id

            join:actor
               type:inner
               on
                  and
                     film_actor.actor_id:actor.actor_id

The above lambda assumes you've got Oracle's Sakila database in your MySQL instance. If you only wish to see its resulting SQL, add a [generate] argument to the above root invocation, and set its value to "true". All specialised slots, dynamically building and executing some SQL towards your database, supports the [generate] argument, allowing you to easily "debug" your SQL statements, and see what they actually do. Below is an example of the SQL created by the above invocation if you add a [generate] argument to it and set its value to boolean true.

select `film`.`title`, `film`.`description`, `actor`.`last_name`, `actor`.`first_name` from `film`
   inner join `film_actor` on `film`.`film_id` = `film_actor`.`film_id`
      inner join `actor` on `film_actor`.`actor_id` = `actor`.`actor_id`
   limit 25

You can recursively join as many levels as you wish, in addition to also supplying multiple join conditions for the same join. An example of the latter can be found below.

sql.read
   generate:true
   limit:-1
   table:table1

      join:table2
         on
            and
               fk1:pk1
               fk2:pk2

The above lambda will result in the following SQL being generated.

select * from 'table1'
   inner join 'table2' on 'table1'.'fk1' = 'table2'.'pk1' and
      'table1'.'fk2' = 'table2'.'pk2'

Joining tables works almost the exact same way as using a [where] argument, allowing you to supply an operator for your join, such as we illustrate below, where we're using the != operator, instead of the (default) equality comparison. However, when you create your [join] segments you can only have columns from tables reference columns from tables, and not add static arguments to your join. If you need static arguments you have to add these into your [where] parts. See the [where] criteria for details about comparison operators. You can also explicitly choose a [type] of join, such as we illustrate below.

sql.read
   limit:-1
   table:table1

      join:table2
         type:inner
         on
            and
               fk1.neq:pk1

The above results in the following SQL.

select * from 'table1' inner join 'table2' on 'table1'.'fk1' != 'table2'.'pk1'

The [type] argument to your [join] arguments, can be "inner", "full", "left" or "right", resulting in the equivalent type of join for your SQL.

Differences between [join] and [where]

There is one crucial semantic difference between a [join] condition and a [where] condition, which is that the library assumes a join is always between two columns, while a where always assume you're always comparing against a static value. This implies that you cannot add static values into your SQL as a part of your [join] condition, while the opposite is true for a [while].

Although this technically doesn't allow you to create any SQL you want to create, it is still more in "the spirit" of SQL as a standard - And you can always add your static conditions into your [where] parts, while adding your table comparisons into your [join] conditions. This allows you to create any result you want to achieve, although technically not any SQL you want to have.

Explicit arguments declarations for joins

Normally you don't need to worry about this, but sometimes you need to explicitly add an argument to your CRUD slot invocations if it has a join part, and you want one of your conditions for your join to be a static value of some sort, and not a comparison to your right hand side table column. This can be accomplished with something such as the following.

sql.read
   generate:true
   table:foo
      join:bar
         on
            and
               foo.field1:bar.field2
               foo.field2:@static-value
   @static-value:static value

Notice the above [@static-value] parts, which becomes a normal argument, instead of trying to join two fields on two table. The above results in the following lambda being generated.

data.read:select * from `foo` inner join `bar` on `foo`.`field1` = `bar`.`field2` and `foo`.`field2` = @static-value limit 25
   @static-value:static value

As you can see in the above result, the @static-value becomes a statically declared condition to your join, and not assumed to be a reference to a field in your right hand side joined table. This is one of those edge cases you normally rarely need, but might be useful on rare occasions. The reasons why this works is because the RHS side of your join condition starts with an @ character, which assumes you are referencing an argument and not a field in your joined table. This only has effects on your [join] parts, implying [xxx.read] slot invocations, since these are the only slots supporting joins. The above is the only exception that allows you to join on static values and not column names. However, as a general rule of thumb, we encourage users to [join] on table columns and add static values into your [where] conditions.

'Namespacing' columns

When you're joining results from multiple tables, it's often required that you specify which table you want some resulting column to be fetched from, to avoid confusing your database as to which column you want to extract, in cases where the same column exists in multiple tables. For such cases, you can simply refer to your table first, and then the column from that table, and separate your entities by a .. You can see an example of this below.

data.connect:sakila
   data.read
      columns

         /*
          * Prefixing result columns with table names.
          */
         film.title
         film.description
         actor.last_name
         actor.first_name

      table:film

         join:film_actor
            type:inner
            on
               and
                  film.film_id:film_actor.film_id

            join:actor
               type:inner
               on
                  and
                     film_actor.actor_id:actor.actor_id

The above will result in the following SQL, if you append the [generate] argument, and set its value to "true".

select
   `film`.`title`,
   `film`.`description`,
   `actor`.`last_name`,
   `actor`.`first_name`
   from `film`
      inner join `film_actor` on `film`.`film_id` = `film_actor`.`film_id`
         inner join `actor` on `film_actor`.`actor_id` = `actor`.`actor_id`
   limit 25

Spacing is not applied to the actual generated SQL result, but have been applied to some of the SQL examples in this documentation to make the SQL more readable. If you supply [as] arguments to your tables, you can also use your alias to reference tables inside of your invocation.

Group by

You can also provide a [group] argument to your lambda, resulting in a "group by" statement injected into the resulting SQL. Below is an example.

sql.read
   table:table1
   limit:-1
   columns
      col1
      count(*)
         as:count
   group
      col1

The above will result in the following SQL.

select col1, count(*) as count from 'table1' group by 'col1'

You can supply multiple group by columns, in addition to "namespacing" your columns with your table names, such as we illustrate below.

sql.read
   table:table1
   limit:-1
   columns
      count(*)
   group
      table1.foo1
      table1.foo2

The above of course results in the following SQL.

select count(*) from 'table1' group by 'table1'.'foo1','table1'.'foo2'

You can of course combine your [group] arguments with [where] arguments, and [join] arguments, allowing you to create complex aggregate results, statistics, joining multiple tables, etc.

How to use [sql.update]

This slot allows you to update one or more records, in a specified [table]. Just like create, it requires one mandatory argument, being [values], implying columns/values you wish to update. This slot also takes an optional [where] argument, which is described further down in this document. Its simplest version can be imagined as follows.

sql.update
   table:table1
   values
      field1:howdy

The above of course will result in the following.

sql.update:update 'table1' set 'field1' = @v0
   @v0:howdy

Notice, if you don't apply a [where] argument, then all records in your table will be updated - Which is highly unlikely what your intentions are. Hence, make sure you apply a [where] argument as you invoke this slot.

How to use [sql.delete]

This slot is for deleting records. Its [where] argument is applied in a similar fashion as the where argument to [sql.select] and [sql.update]. You can find an example further down in this document of how to use [where].

sql.delete
   table:table1
   where
      and
         field1:value1
         field2:value2

The above will produce the following results.

sql.delete:delete from 'table1' where 'field1' = @0 and 'field2' = @1
   @0:value1
   @1:value2

The [where] argument

This argument is common for [sql.update], [sql.delete], and [sql.read], in addition to that a [join] will also be logically parsed much the same way as a [where] argument. The where argument follows a recursive structure, allowing you to supply multiple layers of where criteria, being applied recursively, using some sort of comparison operator, applied to all conditions in the same level. Its most basic usage is as follows.

sql.read
   table:table1
   limit:-1
   where

      and
         field1:howdy

The above would result in the following result.

sql.read:select * from 'table1' where 'field1' = @0
   @0:howdy

To apply multiple [and] criteria, you can simply add them consecutively as follows.

sql.read
   table:table1
   limit:-1
   where

      and
         field1:howdy
         field2:world

The above results in the following.

sql.read:select * from 'table1' where 'field1' = @0 and 'field2' = @1
   @0:howdy
   @1:world

If you exchange the above [and] with [or], the system will use the or operator to separate your arguments, such as the following illustrates.

sql.read
   table:table1
   limit:-1
   where

      or
         field1:howdy
         field2:world

The above results in the following result.

sql.read:select * from 'table1' where 'field1' = @0 or 'field2' = @1
    @0:howdy
    @1:world

You can also nest operators, producing paranthesis, and create complex conditions, such as the following illustrates.

sql.read
   table:table1
   limit:-1
   where

      or
         field1:howdy

         and
            field2:world
            field3:dudes

Which results in the following.

sql.read:select * from 'table1' where 'field1' = @0 or ('field2' = @1 and 'field3' = @2)
   @0:howdy
   @1:world
   @2:dudes

The parent of a list of criteria is deciding which logical operator to separate your conditions with, contrary to traditional languages, where you separate your conditions with the logical operator, and explicitly add paranthesis to group your levels. This might seem a little bit weird in the beginning, but this is a general rule with everything in Hyperlambda, and after a while will feel more natural than the alternative. The reasons for this is to allow for semantically traversing your lambda objects, allowing the computer to logically understand what it does more easily - Among other things. Think of the boolean logical parts of your SQL slots as "grouping your comparisons" or a "logical function invocation" if it helps.

Comparison operators

The project supports the following comparison operators.

  • eq - Equality comparison, equivalent to =
  • neq - Not equality comparison, equivalent to !=
  • mt - More than comparison, equivalent to >
  • lt - Less than comparison, equivalent to <
  • lteq - Less than or equal comparison, equivalent to <=
  • mteq - More than or equal comparison, equivalent to >=
  • like - Like comparison, equivalent to SQL's like comparison
  • in - Special comparison operator, since it requires a list of values, generating an "in" SQL condition

Everywhere you need to compare one field with another, such as in [where] or [join] arguments, you can append a comparison operator to your left hand side column, such as the following illustrates.

sql.read
   table:foo
   where

      and
         field1.neq:xxx

Notice the .neq parts above, and realise how the above will produce the following SQL.

select * from 'foo' where 'field1' != @0

The above [field1.neq] is substituted by the SQL generator to become a != comparison operator on the field1 column versus the xxx argument. The comparison operator is always expected to be the last parts of your "LHS" (Left Hand Side) parts of your criteria - Implying the name of the node.

The [in] comparison operator

This operator is special, in that it doesn't require the caller to supply one value, but rather a list of values, from where the column you compare towards must have a value matching at least one of these values. An example can be found below.

sql.read
   table:table1
   where

      and
         table1.field1.in
            :long:5
            :long:7
            :long:9

The above will generate the following SQL, in addition to returning 3 parameters to the caller.

select * from 'table1' where 'table1'.'field1' in (@0,@1,@2) limit 25

Escaping characters

If you by some freak accident happen to have a column in one of your tables that is named for instance neq, you can escape your column name, by prepending a \ to it. See an example below.

sql.read
   table:table1
   where

      and
         table1.\neq:foo

Notice how the \ character above results in the following SQL.

select * from 'table1' where 'table1'.'neq' = @0 limit 25

As you can see, the \neq is interpreted as a column name, and not a neq operator, and since the equality operator is the default selected if no operator is supplied, the comparison operator becomes =.

You can also escape columns entirely, if you for instance have a column that contains a . in its name, such as we illustrate below.

sql.read
   table:table1
   where

      and
         \foo.bar:bar

Notice how the above lambda will interpret the foo.bar parts as a column name, and not as column "bar" on "foo". You can see the resulting SQL below.

select * from 'table1' where 'foo.bar' = @0 limit 25

Extension comparison operators

You can also extend the existing comparison operators with your own, such as for instance having ltmt resulting in <>, etc. To do this, you'll have to register your comparison operator using the static AddComparisonOperator method on the SqlWhereBuilder class. Below is an example.

SqlWhereBuilder.AddComparisonOperator("ltmt", (builder, args, colNode, escapeChar) => {
    builder.Append(" <> ");
    SqlWhereBuilder.AppendArgs(args, colNode, builder, escapeChar);
});

The above will give you access to use ltmt as a comparison operator, resolving to <> in your SQL.

magic.data.common and meta data

One of the really nice things about this semantic approach to generating SQL is that it allows you to retrieve meta data from your Hyperlambda snippets, asking questions such as for instance "find all files that somehow selects columns from the 'xxx' table" - And for that matter, even dynamically change the table name, using semantic refactoring and replacement concepts. Once you've crossed the initial step into meta data traversal in Hyperlambda, things like this becomes second hand nature with Hyperlambda.

SQL injection attacks

This project protects you automatically against SQL injection attacks, and protect values, and criteria, etc - But you should not allow any potentially insecure clients to dynamically declare which columns to select, and/or field names for your where clauses. It will only protect your values, and not table names or column names against SQL injection attacks.

The project does also not verify that your SQL is possible to execute towards your database, such as verifying that specified tables or columns actually exists. It does its best however, to verify that your Hyperlambda is structured correctly, and that it will create somewhat valid SQL - But you should not assume the SQL the project generates is valid, before you have tested it.

Creating your own data adapter

If you wish to extend Magic to support a custom database type, you can do so using C# for instance. This project contains 4 base classes, which you can inherit from to extend and implement your custom logic.

  • SqlCreateBuilder - Helper class to generate insert SQL statements
  • SqlDeleteBuilder - Helper class to generate delete SQL statements
  • SqlReadBuilder - Helper class to generate select SQL statements
  • SqlUpdateBuilder - Helper class to generate update SQL statements

If you create your own database implementation, you'll need to inherit from the above classes, and override whatever parts of these classes that doesn't by default work as your database type needs it to work. If you wish to do this, you would probably benefit from looking at one of the existing specialised implementations, such as the MySQL, SQL Server or PostgreSQL specific implementations.

Project website for magic.data.common

The source code for this repository can be found at github.com/polterguy/magic.data.common, and you can provide feedback, provide bug reports, etc at the same place.

  • Build status
  • Quality Gate Status
  • Bugs
  • Code Smells
  • Coverage
  • Duplicated Lines (%)
  • Lines of Code
  • Maintainability Rating
  • Reliability Rating
  • Security Rating
  • Technical Debt
  • Vulnerabilities

The projects is copyright of Aista, Ltd 2021 - 2023, and professionally maintained by AINIRO your friendly ChatGPT website chatbot vendor.

Product Compatible and additional computed target framework versions.
.NET net5.0 was computed.  net5.0-windows was computed.  net6.0 was computed.  net6.0-android was computed.  net6.0-ios was computed.  net6.0-maccatalyst was computed.  net6.0-macos was computed.  net6.0-tvos was computed.  net6.0-windows was computed.  net7.0 was computed.  net7.0-android was computed.  net7.0-ios was computed.  net7.0-maccatalyst was computed.  net7.0-macos was computed.  net7.0-tvos was computed.  net7.0-windows was computed.  net8.0 was computed.  net8.0-android was computed.  net8.0-browser was computed.  net8.0-ios was computed.  net8.0-maccatalyst was computed.  net8.0-macos was computed.  net8.0-tvos was computed.  net8.0-windows was computed. 
.NET Core netcoreapp2.0 was computed.  netcoreapp2.1 was computed.  netcoreapp2.2 was computed.  netcoreapp3.0 was computed.  netcoreapp3.1 was computed. 
.NET Standard netstandard2.0 is compatible.  netstandard2.1 was computed. 
.NET Framework net461 was computed.  net462 was computed.  net463 was computed.  net47 was computed.  net471 was computed.  net472 was computed.  net48 was computed.  net481 was computed. 
MonoAndroid monoandroid was computed. 
MonoMac monomac was computed. 
MonoTouch monotouch was computed. 
Tizen tizen40 was computed.  tizen60 was computed. 
Xamarin.iOS xamarinios was computed. 
Xamarin.Mac xamarinmac was computed. 
Xamarin.TVOS xamarintvos was computed. 
Xamarin.WatchOS xamarinwatchos was computed. 
Compatible target framework(s)
Included target framework(s) (in package)
Learn more about Target Frameworks and .NET Standard.

NuGet packages (7)

Showing the top 5 NuGet packages that depend on magic.data.common:

Package Downloads
magic.lambda.logging

Logging helper slots for Magic, allowing you to inject your own logging implementation, giving you the ability to create log entries from Hyperlambda. To use package go to https://polterguy.github.io

magic.lambda.mssql

Microsoft SQL Server helper slots for Magic, allowing you to easily connect to an SQL Server database, and read, update, delete and insert into your database. To use package go to https://polterguy.github.io

magic.lambda.mysql

MySQL Server helper slots for Magic allowing you to easily connect to a MySQL database, and read, update, delete and insert into your database. To use package go to https://polterguy.github.io

magic.lambda.scheduler

Helper slots for Magic to allow you to dynamically create tasks, either as a part of your custom workflow, and/or as scheduled tasks intended to be executed at some interval, or future time. To use package go to https://polterguy.github.io

magic.lambda.odbc

ODBC data adapter helper slots for Magic allowing you to easily connect to an ODBC database, and execute SQL towards that connection. To use package go to https://polterguy.github.io

GitHub repositories

This package is not used by any popular GitHub repositories.

Version Downloads Last updated
17.1.7 629 1/12/2024
17.1.6 587 1/11/2024
17.1.5 634 1/5/2024
17.0.1 713 1/1/2024
17.0.0 1,167 12/14/2023
16.11.5 1,197 11/12/2023
16.9.0 1,254 10/9/2023
16.7.0 1,884 7/11/2023
16.4.1 1,637 7/2/2023
16.4.0 1,729 6/22/2023
16.3.1 1,827 6/7/2023
16.3.0 1,754 5/28/2023
16.1.9 2,226 4/30/2023
15.10.11 2,186 4/13/2023
15.9.1 2,453 3/27/2023
15.9.0 2,275 3/24/2023
15.8.2 2,397 3/20/2023
15.7.0 2,357 3/6/2023
15.5.0 3,988 1/28/2023
15.2.0 3,144 1/18/2023
15.1.0 3,697 12/28/2022
14.5.7 3,419 12/13/2022
14.5.5 3,553 12/6/2022
14.5.1 3,599 11/23/2022
14.5.0 3,608 11/18/2022
14.4.5 4,208 10/22/2022
14.4.1 4,194 10/22/2022
14.4.0 4,127 10/17/2022
14.3.1 4,904 9/12/2022
14.3.0 3,975 9/10/2022
14.1.3 4,330 8/7/2022
14.1.2 4,055 8/7/2022
14.1.1 3,939 8/7/2022
14.0.14 4,033 7/26/2022
14.0.12 3,976 7/24/2022
14.0.11 3,900 7/23/2022
14.0.10 4,091 7/23/2022
14.0.9 4,107 7/23/2022
14.0.8 4,166 7/17/2022
14.0.5 4,165 7/11/2022
14.0.4 4,229 7/6/2022
14.0.3 4,148 7/2/2022
14.0.2 4,022 7/2/2022
14.0.0 4,273 6/25/2022
13.4.0 6,007 5/31/2022
13.3.4 4,886 5/9/2022
13.3.0 5,828 5/1/2022
13.2.0 4,994 4/21/2022
13.1.0 4,429 4/7/2022
13.0.0 4,245 4/5/2022
11.0.5 4,605 3/2/2022
11.0.4 3,866 2/22/2022
11.0.3 2,362 2/9/2022
11.0.2 2,390 2/6/2022
11.0.1 899 2/5/2022
11.0.0 2,313 2/5/2022
10.0.21 2,281 1/28/2022
10.0.20 2,315 1/27/2022
10.0.19 2,296 1/23/2022
10.0.18 2,245 1/17/2022
10.0.16 2,006 1/14/2022
10.0.15 1,623 12/31/2021
10.0.14 1,618 12/28/2021
10.0.7 2,484 12/22/2021
10.0.5 1,703 12/18/2021
10.0.1 1,801 12/13/2021
9.9.9 1,404 11/29/2021
9.9.3 1,693 11/9/2021
9.9.2 1,328 11/4/2021
9.9.0 1,443 10/30/2021
9.8.9 1,509 10/29/2021
9.8.7 1,403 10/27/2021
9.8.6 1,386 10/27/2021
9.8.5 1,500 10/26/2021
9.8.0 2,149 10/20/2021
9.7.9 1,335 10/19/2021
9.7.5 2,284 10/14/2021
9.7.0 1,634 10/9/2021
9.6.7 1,472 8/30/2021
9.6.6 771 8/14/2021
9.4.0 5,365 6/24/2021
9.2.0 3,788 5/26/2021
9.1.4 2,046 4/21/2021
9.1.1 715 4/15/2021
9.1.0 2,178 4/14/2021
9.0.1 1,428 4/12/2021
9.0.0 1,388 4/5/2021
8.9.9 1,850 3/30/2021
8.9.3 2,388 3/19/2021
8.9.2 1,944 1/29/2021
8.9.1 1,932 1/24/2021
8.9.0 1,881 1/22/2021
8.7.1 2,281 11/15/2020
8.7.0 3,214 11/15/2020
8.6.9 2,105 11/8/2020
8.6.6 2,814 11/2/2020
8.6.0 4,936 10/28/2020
8.5.0 2,797 10/23/2020
8.4.0 6,498 10/13/2020
8.3.1 3,428 10/5/2020
8.3.0 2,066 10/3/2020
8.2.2 2,925 9/26/2020
8.2.1 2,060 9/25/2020
8.2.0 2,162 9/25/2020
8.1.19 956 9/21/2020
8.1.18 6,641 9/13/2020
8.1.17 7,592 9/13/2020
8.1.16 951 9/13/2020
8.1.15 3,326 9/12/2020
8.1.11 3,466 9/11/2020
8.1.10 2,193 9/6/2020
8.1.9 2,226 9/3/2020
8.1.8 2,182 9/2/2020
8.1.7 1,985 8/28/2020
8.1.4 2,108 8/25/2020
8.1.3 2,109 8/18/2020
8.1.2 2,119 8/16/2020
8.1.1 2,237 8/15/2020
8.1.0 990 8/15/2020
8.0.1 4,105 8/7/2020
8.0.0 2,068 8/7/2020
7.0.1 995 6/28/2020
7.0.0 3,299 6/28/2020
5.0.0 9,042 2/25/2020
4.0.4 8,771 1/27/2020
4.0.3 2,161 1/27/2020
4.0.2 2,197 1/16/2020
4.0.1 2,182 1/11/2020
4.0.0 2,296 1/5/2020
3.1.0 7,432 11/10/2019
3.0.1 5,685 10/28/2019
3.0.0 5,043 10/23/2019
2.0.2 1,744 10/21/2019
2.0.1 10,682 10/15/2019
2.0.0 2,424 10/13/2019
1.1.9 2,254 10/11/2019
1.1.8 2,344 10/10/2019
1.1.7 1,458 10/9/2019
1.1.6 1,002 10/9/2019
1.1.5 944 10/8/2019
1.1.4 2,985 10/6/2019
1.1.3 1,471 10/5/2019
1.1.2 1,300 10/5/2019
1.0.1 985 9/30/2019
1.0.0 1,275 9/26/2019