steel soft: an erp application foe steel company

62
STEEL SOFT: AN ERP APPLICATION FOE STEEL COMPANY Steel soft: An ERP Application for steel company is powerful, flexible, and easy to use and is designed and developed to deliver real conceivable benefits to Steel companys.Steel company Management System is designed to cover a wide range of Steel company administration and management processes. It is an integrated end-to-end Steel company Management System that provides relevant information across the Steel company to support effective decision making for patient care, Steel company administration and critical financial accounting, in a seamless flow.Steel company Management System is a software product suite designed to improve the quality and management of Steel company management in the areas of clinical process analysis and activity-based costing. Steel company Management System enables to develop yorganization and improve its effectiveness and quality of work. Managing the key processes efficiently is critical to the success of the Steel company helps manage processesThe Sales and Inventory System objective is to manage a steel company and its marketing which is a very huge task. Through the considerable amount of time used by the automated system, the end users day to day job of managing the shop will be reduced. The system will take care of all the customers in a quick time. The users will consume less amount of time when compared to manual paper work through the automated system. The system will take care of all the sales activities done in a shop.

Upload: b-u

Post on 26-Nov-2023

0 views

Category:

Documents


0 download

TRANSCRIPT

STEEL SOFT: AN ERP APPLICATION FOE STEEL COMPANY

Steel soft: An ERP Application for steel company is powerful, flexible, and easy to use

and is designed and developed to deliver real conceivable benefits to Steel companys.Steel

company Management System is designed to cover a wide range of Steel company

administration and management processes. It is an integrated end-to-end Steel company

Management System that provides relevant information across the Steel company to support

effective decision making for patient care, Steel company administration and critical financial

accounting, in a seamless flow.Steel company Management System is a software product suite

designed to improve the quality and management of Steel company management in the areas of

clinical process analysis and activity-based costing. Steel company Management System enables

to develop yorganization and improve its effectiveness and quality of work. Managing the key

processes efficiently is critical to the success of the Steel company helps manage processesThe

Sales and Inventory System objective is to manage a steel company and its marketing which is a

very huge task. Through the considerable amount of time used by the automated system, the end

users day to day job of managing the shop will be reduced. The system will take care of all the

customers in a quick time. The users will consume less amount of time when compared to

manual paper work through the automated system. The system will take care of all the sales

activities done in a shop.

Modules:

Customers

Employees

Product

Stock

Sales

Sales Return

Purchase

Purchase Return

Bill

Report

Modules Description:

Customer Details:

This module contains the information about the customers of different items to various

categories are maintained along with their details. Their details are stored in the database and

retrieved whenever needed.

Employees Details:

The employee module contains the information about the employees who are working in

a Departmental Shop. Their details such as employee id, employee name, and their contacts are

stored in the database and retrieved whenever needed.

Product Details:

The Product module contains the information about the products available a

Departmental Shop. Their details such as Product Id, Product name, and their catagories are

stored in the database and retrieved whenever needed.

Stock Details:

The Stock module contains the information about the products available a Departmental

Shop. Their details such as Stock Id, Product name, and their categories are stored in the

database and retrieved whenever needed.

Sales Details:

This module contains the information about the sales of different items to various

customer are maintained along with their details. Their details are stored in the database and

retrieved whenever needed.

Sales Return Details:

This module contains the information of complained products . This products which are

returned by the customer to the store . Their details are stored in these database and retrieved

whenever needed.

Purchase Details:

This module contains the information about the purchase of an item from different

suppliers. Their purchase details are stored in the database and also can be retrieved.

Purchase Return Details:

This module contains the information of complained products . This products are

returned to the supplier . Their details are stored in these database and retrieved whenever

needed.

Bill Details:

This module contains the billing information of products. Once the product has been sold

to the customer billing is generated along with the product information.

Report:

Reports are generated by the system which has been stored in the centralized database

and the types of reports like photos details, product details, employee details and so.

2.SYSTEM STUDY

Existing System:

In current system the information is very difficult to retrieve and to find particular

information Manual calculations are error prone and take a lot of time this may result in incorrect

information. A difficult task as information is difficult to collect from various registers. The

information generated by various transactions takes time and efforts to be stored at right place.

Proposed System:

Data storing is easier. Paper work will be reduced and the user spends more time on

monitoring the progress. The system is user friendly and easy to use. All the important data’s

will be stored in the database and it avoids any miscalculation. This system is helpful to

computerize the scheduled events and also very helpful in calculating the bill without any

miscalculation. The reports can be checked depending on month/year.

3.SYSTEM SPECIFICATION

3.1 HARDWARE SPECIFICATION

Processor: Intel dual core or above

Processor Speed: 1.0GHZ or above

RAM: 1 GB RAM or above

Hard Disk: 20 GB hard disk or above

3.2 SOFTWARE SPECIFICATION:

Language: Asp.net 2010.

Database: Microsoft SQL server 2008

3.2.1 ABOUT THE FRONT END

ASP.NET is part of the .NET framework. ASP.NET programs are centralized

applications hosted on one or more Web servers that respond dynamically to client requests. The

responses are dynamic because ASP.NET intercepts requests for pages with a specific extension (.aspx

or .ascx) and hands off the responsibility for answering those requests to just-in-time (JIT) compiled code

files that can build a response “on-the-fly.”

ASP.NET deals specifically with configuration (web. comfit and machine. config) files,

Web Services (ASMX) files, and Web Forms (ASPX) files. The server doesn’t “serve” any of these file

types—it returns the appropriate content type to the client. The configuration file types contain

initialization and settings for a specific application or portion of an application. Another configuration

file, called machine.web, contains machine-level initialization and settings. The server ignores requests

for web files, because serving them might constitute a security breach.

Client requests for these file types cause the server to load, parse, and execute code to

return a dynamic response. For Web Forms, the response usually consists of HTML or WML. Web Forms

maintain state by round-tripping user interface and other persistent values between the client and the

server automatically for each request.

A request for a Web Form can use View State, Session State, or Application State to

maintain values between requests. Both Web Forms and Web Services requests can take advantage of

ASP. Net’s integrated security and data access through ADO.NET, and can run code that uses system

services to construct the response. So the major difference between a static request and a dynamic request

is that a typical Web request references a static file. The server reads the file and responds with the

contents of the requested file.

ASP.NET uses .NET languages. ASP.NET code exists in multithreaded JIT compiled

DLL assemblies, which can be loaded on demand. Once loaded, the ASP.NET DLLs can service multiple

requests from a single in-memory copy.

ASP.NET supports all the .NET languages (currently C#, C++, VB.NET, and JScript, but

there are well over 20 different languages in development for .NET), so you will eventually be able to

write Web applications in your choice of almost any modern programming language. In addition to huge

increases in speed and power, ASP.NET provides substantial development improvements, like seamless

server-to-client debugging, automatic validation of form data.

Fig2. Interoperability

ADO.NET

ADO.NET provides a set of classes which a script can use to interact with databases.

Scripts can create instances of ADO.NET data classes and access their properties and methods. A set of

classes which work with a specific type of database is known as a .NET Data Provider. ADO.NET

comes with two Data Providers, the SQL Server.NET Data Provider (which provides optimised access for

Microsoft SQL Server databases) and the OLEDB.NET Data Provider, which works with a range of

databases. The main ADO.NET OLEDB data access classes are OLEDBConnection, OLEDBCommand,

OLEDBDataReader and OLEDBDataAdapter.

FEATURES OF SQL SERVER 2000

The OLAP Services feature available in SQL Server version 7.0 is now called SQL

Server 2000 Analysis Services. The term OLAP Services has been replaced with the term

Analysis Services. Analysis Services also includes a new data mining component. The

Repository component available in SQL Server version 7.0 is now called Microsoft SQL Server

2000 Meta Data Services. References to the component now use the term Meta Data Services.

The term repository is used only in reference to the repository engine within Meta Data Services

SQL-SERVER database consist of six type of objects,

They are,

1. TABLE

2. QUERY

3. FORM

4. REPORT

5. MACRO

TABLE:

A database is a collection of data about a specific topic.

VIEWS OF TABLE:

We can work with a table in two types,

1. Design View

2. Datasheet View

Design View

To build or modify the structure of a table we work in the table design view. We can

specify what kind of data will be hold.

Datasheet View

To add, edit or analyses the data itself we work in tables datasheet view mode.

QUERY:

A query is a question that has to be asked the data. Access gathers data that answers the

question from one or more table. The data that make up the answer is either dynaset (if you edit

it) or a snapshot(it cannot be edited).Each time we run query, we get latest information in the

dynaset.Access either displays the dynaset or snapshot for us to view or perform an action on

it ,such as deleting or updating.

FORMS:

A form is used to view and edit information in the database record by record .A form

displays only the information we want to see in the way we want to see it. Forms use the familiar

controls such as textboxes and checkboxes. This makes viewing and entering data easy.

views of Form:

We can work with forms in several primarily there are two views,

They are,

1. Design View

2. Form View

Design View

To build or modify the structure of a form, we work in forms design view. We can add

control to the form that are bound to fields in a table or query, includes textboxes, option buttons,

graphs and pictures.

Form View

The form view which display the whole design of the form.

REPORT:

A report is used to vies and print information from the database. The report can ground

records into many levels and compute totals and average by checking values from many records

at once. Also the report is attractive and distinctive because we have control over the size and

appearance of it.

MACRO :

A macro is a set of actions. Each action in macros does something. Such as opening a

form or printing a report .We write macros to automate the common tasks the work easy and

save the time.

4.SYSTEM DESIGN

LEVEL 0:

Request

Inventory

Response

Admin Steel management

LEVEL: 1

Request

Admin

Response

I insert

Customer

View

Update

Employees

View

Update Product

View

update Stock

View

update

View Purchase

Admin Login

Customer Details

Employees Details

Product Details

Stock Details

Purchase Details

Report

4.2 TABLE DESIGN

DATABASE NAME: Inventory

TABLE 1: Customer

FIELD NAME DATA TYPE CONSTRAINTS

Id int PrimaryKey

Name NVarchar(Max) Not Null

Address NVarchar(Max) Not Null

Email Nvarchar(Max) Not Null

Mobile num int Not Null

TABLE 2: Employees

FIELD NAME DATA TYPE CONSTRAINTS

Id int PrimaryKey

Emp Name NVarchar(Max) Not Null

Address NVarchar(Max) Not Null

Department NVarchar(Max) Not Null

Email NVarchar(Max) Not Null

Mobile num int Not Null

TABLE : Product

FIELD NAME DATA TYPE CONSTRAINTS

Product Id int PrimaryKey

Product Name NVarchar(Max) Not Null

Product type NVarchar(Max) Not Null

Amount Money Not Null

Quantity int Not Null

TABLE : Stock

FIELD NAME DATA TYPE CONSTRAINTS

Product Id int Foreign key

Stock total int Not Null

TABLE : Sales

FIELD NAME DATA TYPE CONSTRAINTS

Product Id int Foreign key

Customer id int Foreign key

Sales date datetime Not Null

Amount Money Not Null

Quantity int Not Null

TABLE : purchase

FIELD NAME DATA TYPE CONSTRAINTS

Product Id int Foreign key

Customer id int Foreign key

puchase date datetime Not Null

Amount Money Not Null

Purchase Quantity int Not Null

4.3 INPUT DESIGN

The input design is the link between the information system and the user. It comprises the

developing specification and procedures for data preparation and those steps are necessary to put

transaction data in to a usable form for processing can be achieved by inspecting the computer to

read data from a written or printed document or it can occur by having people keying the data

directly into the system. The design of input focuses on controlling the amount of input required,

controlling the errors, avoiding delay, avoiding extra steps and keeping the process simple. The

input is designed in such a way so that it provides security and ease of use with retaining the

privacy. Input Design considered the following things:

What data should be given as input?

How the data should be arranged or coded?

The dialog to guide the operating personnel in providing input.

Methods for preparing input validations and steps to follow when error occur.

OBJECTIVES

1. Input Design is the process of converting a user-oriented description of the input into a

computer-based system. This design is important to avoid errors in the data input process and

show the correct direction to the management for getting correct information from the

computerized system.

2. It is achieved by creating user-friendly screens for the data entry to handle large volume of

data. The goal of designing input is to make data entry easier and to be free from errors. The data

entry screen is designed in such a way that all the data manipulates can be performed. It also

provides record viewing facilities.

3. When the data is entered it will check for its validity. Data can be entered with the help of

screens. Appropriate messages are provided as when needed so that the user

will not be in maize of instant. Thus the objective of input design is to create an input layout that is easy to follow

4.4 OUTPUT DESIGN

A quality output is one, which meets the requirements of the end user and presents the

information clearly. In any system results of processing are communicated to the users and to

other system through outputs. In output design it is determined how the information is to be

displaced for immediate need and also the hard copy output. It is the most important and direct

source information to the user. Efficient and intelligent output design improves the system’s

relationship to help user decision-making.

1. Designing computer output should proceed in an organized, well thought out manner; the right

output must be developed while ensuring that each output element is designed so that people will

find the system can use easily and effectively. When analysis design computer output, they

should Identify the specific output that is needed to meet the requirements.

2.Select methods for presenting information.

3.Create document, report, or other formats that contain information produced by the system.

The output form of an information system should accomplish one or more of the following

objectives.

Convey information about past activities, current status or projections of the

Future.

Signal important events, opportunities, problems, or warnings.

Trigger an action.

Confirm an action.

5. TESTING AND IMPLEMENTATION PHASE

SYSTEM TEST

System testing ensures that the entire integrated software system meets requirements. It

tests a configuration to ensure known and predictable results. An example of system testing is the

configuration oriented system integration test. System testing is based on process descriptions

and flows, emphasizing pre-driven process links and integration points. Unit testing involves the

design of test cases that validate that the internal program logic is functioning properly, and that

program inputs produce valid outputs. All decision branches and internal code flow should be

validated. It is the testing of individual software units of the application .it is done after the

completion of an individual unit before integration. This is a structural testing, that relies on

knowledge of its construction and is invasive. Unit tests perform basic tests at component level

and test a specific business process, application, and/or system configuration. Unit tests ensure

that each unique path of a business process performs accurately to the documented specifications

and contains clearly defined inputs and expected results.

INTEGRATION TESTING

Integration tests are designed to test integrated software components to determine if they

actually run as one program. Testing is event driven and is more concerned with the basic

outcome of screens or fields. Integration tests demonstrate that although the components were

individually satisfaction, as shown by successfully unit testing, the combination of components is

correct and consistent. Integration testing is specifically aimed at exposing the problems that

arise from the combination of components.

FUNCTIONAL TEST

Functional tests provide systematic demonstrations that functions tested are available as

specified by the business and technical requirements, system documentation, and user manuals.

Functional testing is centered on the following items:

Valid Input : identified classes of valid input must be accepted.

Invalid Input : identified classes of invalid input must be rejected.

Functions : identified functions must be exercised.

Output : identified classes of application outputs must be exercised.

Systems/Procedures: interfacing systems or procedures must be invoked.

Organization and preparation of functional tests is focused on requirements, key

functions, or special test cases. In addition, systematic coverage pertaining to identify

Business process flows; data fields, predefined processes, and successive processes

must be considered for testing. Before functional testing is complete, additional tests

are identified and the effective value of current tests is determined.

WHITE BOX TESTING

White Box Testing is a testing in which in which the software tester has knowledge of the

inner workings, structure and language of the software, or at least its purpose. It is purpose. It is

used to test areas that cannot be reached from a black box level.

BLACK BOX TESTING

Black Box Testing is testing the software without any knowledge of the inner workings,

structure or language of the module being tested. Black box tests, as most other kinds of tests,

must be written from a definitive source document, such as specification or requirements

document, such as specification or requirements document. It is a testing in which the software

under test is treated, as a black box .you cannot “see” into it. The test provides inputs and

responds to outputs without considering how the software works.

ACCEPTANCE TESTING

After the system test has corrected all or most defects, the system will be delivered to the

user or customer for acceptance testing. Acceptance testing is basically done by the user or

customer although other stakeholders may be involved as well. The goaof acceptance testing is

to establish confidence in the system. Acceptance testing is most often focused on a validation

type testing.

ALPHA TESTING (VERIFICATION TESTING)

This test takes place at the developer’s site. Developers observe the users and note

problems. Alpha testing is testing of an application when development is about to complete.

Minor design changes can still be made as a result of alpha testing. Alpha testing is final testing

before the software is released to the general public.

BETA TESTING (VALIDATION TESTING)

It is also known as field testing. It takes place at customer’s site. It sends the system to

users who install it and use it under real-world working conditions. The goal of beta testing is to

place your application in the hands of real users outside of your own engineering team to

discover any flaws or issues.

5.2 SYSTEM IMPLEMENTATION

The Microsoft .NET Framework is a software technology that is available with several

Microsoft Windows operating systems. It includes a large library of pre-coded solutions to

common programming problems and a virtual machine that manages the execution of programs

written specifically for the framework. The .NET Framework is a key Microsoft offering and is

intended to be used by most new applications created for the Windows platform. The pre-coded

solutions that form the framework's Base Class Library cover a large range of programming

needs in a number of areas, including user interface, data access, database connectivity,

cryptography, web application development, numeric algorithms, and network communications.

The class library is used by programmers, who combine it with their own code to produce

applications.

Programs written for the .NET Framework execute in a software environment that

manages the program's runtime requirements. Also part of the .NET Framework, this runtime

environment is known as the Common Language Runtime (CLR). The CLR provides the

appearance of an application virtual machine so that programmers need not consider the

capabilities of the specific CPU that will execute the program. The CLR also provides other

important services such as security, memory management, and exception handling. The class

library and the CLR together compose the .NET Framework.

Because interaction between new and older applications is commonly required, the .NET

Framework provides means to access functionality that is implemented in programs that execute

outside the .NET environment. Access to COM components is provided in the

System.Runtime.InteropServices and System.EnterpriseServices namespaces of the framework;

access to other functionality is provided using the P/Invoke feature.

Common Runtime Engine 

The Common Language Runtime (CLR) is the virtual machine component of the .NET

framework. All .NET programs execute under the supervision of the CLR, guaranteeing certain

properties and behaviors in the areas of memory management, security, and exception handling.

Base Class Library 

The Base Class Library (BCL), part of the Framework Class Library (FCL), is a library

of functionality available to all languages using the .NET Framework. The BCL provides classes

which encapsulate a number of common functions, including file reading and writing, graphic

rendering, database interaction and XML document manipulation.

Simplified Deployment 

Installation of computer software must be carefully managed to ensure that it does not

interfere with previously installed software, and that it conforms to security requirements.

The .NET framework includes design features and tools that help address these requirements.

Security

The design is meant to address some of the vulnerabilities, such as buffer overflows, that

have been exploited by malicious software. Additionally, .NET provides a common security

model for all applications.

Portability 

The design of the .NET Framework allows it to theoretically be platform agnostic, and

thus cross-platform compatible. That is, a program written to use the framework should run

without change on any type of system for which the framework is implemented. Microsoft's

commercial implementations of the framework cover Windows, Windows CE, and the Xbox

360. In addition, Microsoft submits the specifications for the Common Language Infrastructure

(which includes the core class libraries, Common Type System, and the Common Intermediate

Language), the C# language, and the C++/CLI language to both ECMA and the ISO, making

them available as open standards. This makes it possible for third parties to create compatible

implementations of the framework and its languages on other platforms.

The .NET Framework CLR frees the developer from the burden of managing memory

(allocating and freeing up when done); instead it does the memory management itself. To this

end, the memory allocated to instantiations of .NET types (objects) is done contiguously from

the managed heap, a pool of memory managed by the CLR. As long as there exists a reference to

an object, which might be either a direct reference to an object or via a graph of objects, the

object is considered to be in use by the CLR. When there is no reference to an object, and it

cannot be reached or used, it becomes garbage. However, it still holds on to the memory

allocated to it. .NET Framework includes a garbage collector which runs periodically, on a

separate thread from the application's thread, that enumerates all the unusable objects and

reclaims the memory allocated to them.

The .NET Garbage Collector (GC) is a non-deterministic, compacting, mark-and-sweep

garbage collector. The GC runs only when a certain amount of memory has been used or there is

enough pressure for memory on the system. Since it is not guaranteed when the conditions to

reclaim memory are reached, the GC runs are non-deterministic. Each .NET application has a set

of roots, which are pointers to objects on the managed heap (managed objects). These include

references to static objects and objects defined as local variables or method parameters currently

in scope, as well as objects referred to by CPU registers. When the GC runs, it pauses the

application, and for each object referred to in the root, it recursively enumerates all the objects

reachable from the root objects and marks them as reachable. It uses .NET metadata and

reflection to discover the objects encapsulated by an object, and then recursively walk them. It

then enumerates all the objects on the heap (which were initially allocated contiguously) using

reflection. All objects not marked as reachable are garbage. This is the mark phase. Since the

memory held by garbage is not of any consequence, it is considered free space. However, this

leaves chunks of free space between objects which were initially contiguous. The objects are

then compacted together, by using memory to copy them over to the free space to make them

contiguous again. Any reference to an object invalidated by moving the object is updated to

reflect the new location by the GC. The application is resumed after the garbage collection is

over.

The GC used by .NET Framework is actually generational. Objects are assigned a

generation; newly created objects belong to Generation 0. The objects that survive a garbage

collection are tagged as Generation 1, and the Generation 1 objects that survive another

collection are Generation 2 objects. The .NET Framework uses up to Generation 2 objects.

Higher generation objects are garbage collected less frequently than lower generation objects.

This helps increase the efficiency of garbage collection, as older objects tend to have a larger

lifetime than newer objects. Thus, by removing older (and thus more likely to survive a

collection) objects from the scope of a collection run, fewer objects need to be checked and

compacted.

ACTIVE SERVER PAGES.NET

ASP.NET is a programming framework built on the common language runtime that can be used

on a server to build powerful Web applications. ASP.NET offers several important advantages

over previous Web development models:

Enhanced Performance. ASP.NET is compiled common language runtime code running

on the server. Unlike its interpreted predecessors, ASP.NET can take advantage of early

binding, just-in-time compilation, native optimization, and caching services right out of

the box. This amounts to dramatically better performance before you ever write a line of

code.

World-Class Tool Support. The ASP.NET framework is complemented by a rich

toolbox and designer in the Visual Studio integrated development environment.

WYSIWYG editing, drag-and-drop server controls, and automatic deployment are just a

few of the features this powerful tool provides.

Power and Flexibility. Because ASP.NET is based on the common language runtime,

the power and flexibility of that entire platform is available to Web application

developers. The .NET Framework class library, Messaging, and Data Access solutions

are all seamlessly accessible from the Web. ASP.NET is also language-independent, so

you can choose the language that best applies to your application or partition your

application across many languages. Further, common language runtime interoperability

guarantees that your existing investment in COM-based development is preserved when

migrating to ASP.NET.

Simplicity. ASP.NET makes it easy to perform common tasks, from simple form

submission and client authentication to deployment and site configuration. For example,

the ASP.NET page framework allows you to build user interfaces that cleanly separate

application logic from presentation code and to handle events in a simple, Visual Basic -

like forms processing model. Additionally, the common language runtime simplifies

development, with managed code services such as automatic reference counting and

garbage collection.

Manageability. ASP.NET employs a text-based, hierarchical configuration system,

which simplifies applying settings to your server environment and Web applications.

Because configuration information is stored as plain text, new settings may be applied

without the aid of local administration tools. This "zero local administration" philosophy

extends to deploying ASP.NET Framework applications as well. An ASP.NET

Framework application is deployed to a server simply by copying the necessary files to

the server. No server restart is required, even to deploy or replace running compiled code.

Scalability and Availability. ASP.NET has been designed with scalability in mind, with

features specifically tailored to improve performance in clustered and multiprocessor

environments. Further, processes are closely monitored and managed by the ASP.NET

runtime, so that if one misbehaves (leaks, deadlocks), a new process can be created in its

place, which helps keep your application constantly available to handle requests.

Customizability and Extensibility. ASP.NET delivers a well-factored architecture that

allows developers to "plug-in" their code at the appropriate level. In fact, it is possible to

extend or replace any subcomponent of the ASP.NET runtime with your own custom-

written component. Implementing custom authentication or state services has never been

easier.

Security. With built in Windows authentication and per-application configuration, you

can be assured that your applications are secure.

LANGUAGE SUPPORT

The Microsoft .NET Platform currently offers built-in support for three languages: C#,

Visual Basic, and Java Script.

WHAT IS ASP.NET WEB FORMS?

The ASP.NET Web Forms page framework is a scalable common language runtime

programming model that can be used on the server to dynamically generate Web pages. Intended

as a logical evolution of ASP (ASP.NET provides syntax compatibility with existing pages), the

ASP.NET Web Forms framework has been specifically designed to address a number of key

deficiencies in the previous model. In particular, it provides:

The ability to create and use reusable UI controls that can encapsulate common

functionality and thus reduce the amount of code that a page developer has to

write.

The ability for developers to cleanly structure their page logic in an orderly

fashion (not "spaghetti code").

The ability for development tools to provide strong WYSIWYG design support

for pages (existing ASP code is opaque to tools).

ASP.NET Web Forms pages are text files with an .aspx file name extension. They can be

deployed throughout an IIS virtual root directory tree. When a browser client requests .aspx

resources, the ASP.NET runtime parses and compiles the target file into a .NET Framework

class. This class can then be used to dynamically process incoming requests. (Note that the .aspx

file is compiled only the first time it is accessed; the compiled type instance is then reused across

multiple requests).

An ASP.NET page can be created simply by taking an existing HTML file and changing its file

name extension to .aspx (no modification of code is required). For example, the following

sample demonstrates a simple HTML page that collects a user's name and category preference

and then performs a form post back to the originating page when a button is clicked:

ASP.NET provides syntax compatibility with existing ASP pages. This includes support for <%

%> code render blocks that can be intermixed with HTML content within an .aspx file. These

code blocks execute in a top-down manner at page render time.

ADO.NET OVERVIEW

ADO.NET is an evolution of the ADO data access model that directly addresses user

requirements for developing scalable applications. It was designed specifically for the web with

scalability, statelessness, and XML in mind.

ADO.NET uses some ADO objects, such as the Connection and Command objects, and also

introduces new objects. Key new ADO.NET objects include the Dataset, Data Reader, and

Data Adapter.

The important distinction between this evolved stage of ADO.NET and previous data

architectures is that there exists an object -- the DataSet -- that is separate and distinct from any

data stores. Because of that, the DataSet functions as a standalone entity. You can think of the

DataSet as an always disconnected recordset that knows nothing about the source or destination

of the data it contains. Inside a DataSet, much like in a database, there are tables, columns,

relationships, constraints, views, and so forth. A DataAdapter is the object that connects to the

database to fill the DataSet. Then, it connects back to the database to update the data there,

based on operations performed while the DataSet held the data. In the past, data processing has

been primarily connection-based. Now, in an effort to make multi-tiered apps more efficient,

data processing is turning to a message-based approach that revolves around chunks of

information. At the center of this approach is the DataAdapter, which provides a bridge to

retrieve and save data between a DataSet and its source data store. It accomplishes this by means

of requests to the appropriate SQL commands made against the data store.

The XML-based DataSet object provides a consistent programming model that works with all

models of data storage: flat, relational, and hierarchical. It does this by having no 'knowledge' of

the source of its data, and by representing the data that it holds as collections and data types. No

matter what the source of the data within the DataSet is, it is manipulated through the same set

of standard APIs exposed through the DataSet and its subordinate objects.

While the DataSet has no knowledge of the source of its data, the managed provider has

detailed and specific information. The role of the managed provider is to connect, fill, and persist

the DataSet to and from data stores. The OLE DB and SQL Server .NET Data Providers

(System.Data.OleDb and System.Data.SqlClient) that are part of the .Net Framework provide

four basic objects: the Command, Connection, DataReader and DataAdapter. In the

remaining sections of this document, we'll walk through each part of the DataSet and the OLE

DB/SQL Server .NET Data Providers explaining what they are, and how to program against

them.

The following sections will introduce you to some objects that have evolved, and some that are

new. These objects are:

Connections. For connection to and managing transactions against a database.

Commands. For issuing SQL commands against a database.

DataReaders. For reading a forward-only stream of data records from a SQL Server data

source.

DataSet. For storing, Remoting and programming against flat data, XML data and

relational data.

DataAdapters. For pushing data into a DataSet, and reconciling data against a database.

When dealing with connections to a database, there are two different options: SQL Server .NET

Data Provider (System.Data.SqlClient) and OLE DB .NET Data Provider (System.Data.OleDb).

In these samples we will use the SQL Server .NET Data Provider. These are written to talk

directly to Microsoft SQL Server. The OLE DB .NET Data Provider is used to talk to any OLE

DB provider (as it uses OLE DB underneath).

Connections:

Connections are used to 'talk to' databases, and are represented by provider-specific classes such

as SqlConnection. Commands travel over connections and resultsets are returned in the form of

streams which can be read by a DataReader object, or pushed into a DataSet object.

Commands:

Commands contain the information that is submitted to a database, and are represented by

provider-specific classes such as SqlCommand. A command can be a stored procedure call, an

UPDATE statement, or a statement that returns results. You can also use input and output

parameters, and return values as part of your command syntax. The example below shows how

to issue an INSERT statement against the Northwind database.

DataReaders:

The Data Reader object is somewhat synonymous with a read-only/forward-only cursor

over data. The DataReader API supports flat as well as hierarchical data. A DataReader object

is returned after executing a command against a database. The format of the returned

DataReader object is different from a recordset. For example, you might use the DataReader to

show the results of a search list in a web page.

DATASETS AND DATAADAPTERS:

DataSets

The Dataset object is similar to the ADO Recordset object, but more powerful, and with

one other important distinction: the DataSet is always disconnected. The DataSet object

represents a cache of data, with database-like structures such as tables, columns, relationships,

and constraints. However, though a DataSet can and does behave much like a database, it is

important to remember that DataSet objects do not interact directly with databases, or other

source data. This allows the developer to work with a programming model that is always

consistent, regardless of where the source data resides. Data coming from a database, an XML

file, from code, or user input can all be placed into DataSet objects. Then, as changes are made

to the DataSet they can be tracked and verified before updating the source data. The

GetChanges method of the DataSet object actually creates a second DatSet that contains only

the changes to the data. This DataSet is then used by a DataAdapter (or other objects) to update

the original data source.

The DataSet has many XML characteristics, including the ability to produce and

consume XML data and XML schemas. XML schemas can be used to describe schemas

interchanged via WebServices. In fact, a DataSet with a schema can actually be compiled for

type safety and statement completion.

DATAADAPTERS (OLEDB/SQL)

The DataAdapter object works as a bridge between the DataSet and the source data.

Using the provider-specific SqlDataAdapter (along with its associated SqlCommand and

SqlConnection) can increase overall performance when working with a Microsoft SQL Server

databases. For other OLE DB-supported databases, you would use the OleDbDataAdapter

object and its associated OleDbCommand and OleDbConnection objects.

The DataAdapter object uses commands to update the data source after changes have been

made to the DataSet. Using the Fill method of the DataAdapter calls the SELECT command;

using the Update method calls the INSERT, UPDATE or DELETE command for each changed

row. You can explicitly set these commands in order to control the statements used at runtime to

resolve changes, including the use of stored procedures. For ad-hoc scenarios, a

CommandBuilder object can generate these at run-time based upon a select statement.

However, this run-time generation requires an extra round-trip to the server in order to gather

required metadata, so explicitly providing the INSERT, UPDATE, and DELETE commands at

design time will result in better run-time performance.

1. ADO.NET is the next evolution of ADO for the .Net Framework.

2. ADO.NET was created with n-Tier, statelessness and XML in the forefront. Two new

objects, the DataSet and DataAdapter, are provided for these scenarios.

3. ADO.NET can be used to get data from a stream, or to store data in a cache for updates.

4. There is a lot more information about ADO.NET in the documentation.

5. Remember, you can execute a command directly against the database in order to do

inserts, updates, and deletes. You don't need to first put data into a DataSet in order to

insert, update, or delete it.

Also, you can use a DataSet to bind to the data, move through the data, and navigate data

relationships

A database management, or DBMS, gives the user access to their data and helps them

transform the data into information. Such database management systems include dBase, paradox,

IMS, SQL Server and SQL Server. These systems allow users to create, update and extract

information from their database.

A database is a structured collection of data. Data refers to the characteristics of people,

things and events. SQL Server stores each data item in its own fields. In SQL Server, the fields

relating to a particular person, thing or event are bundled together to form a single complete unit

of data, called a record (it can also be referred to as raw or an occurrence). Each record is made

up of a number of fields. No two fields in a record can have the same field name.

During an SQL Server Database design project, the analysis of your business needs

identifies all the fields or attributes of interest. If your business needs change over time, you

define any additional fields or change the definition of existing fields.

SQL SERVER TABLES

SQL Server stores records relating to each other in a table. Different tables are created

for the various groups of information. Related tables are grouped together to form a database.

PRIMARY KEY

Every table in SQL Server has a field or a combination of fields that uniquely identifies

each record in the table. The Unique identifier is called the Primary Key, or simply the Key.

The primary key provides the means to distinguish one record from all other in a table. It allows

the user and the database system to identify, locate and refer to one particular record in the

database.

RELATIONAL DATABASE

Sometimes all the information of interest to a business operation can be stored in one

table. SQL Server makes it very easy to link the data in multiple tables. Matching an employee

to the department in which they work is one example. This is what makes SQL Server a

relational database management system, or RDBMS. It stores data in two or more tables and

enables you to define relationships between the table and enables you to define relationships

between the tables.

FOREIGN KEY

When a field is one table matches the primary key of another field is referred to as a

foreign key. A foreign key is a field or a group of fields in one table whose values match those

of the primary key of another table.

REFERENTIAL INTEGRITY

Not only does SQL Server allow you to link multiple tables, it also maintains consistency

between them. Ensuring that the data among related tables is correctly matched is referred to as

maintaining referential integrity.

DATA ABSTRACTION

A major purpose of a database system is to provide users with an abstract view of the

data. This system hides certain details of how the data is stored and maintained. Data abstraction

is divided into three levels.

Physical level: This is the lowest level of abstraction at which one describes how the data are

actually stored.

Conceptual Level: At this level of database abstraction all the attributed and what data are

actually stored is described and entries and relationship among them.

View level: This is the highest level of abstraction at which one describes only part of the

database.

ADVANTAGES OF RDBMS

Redundancy can be avoided

Inconsistency can be eliminated

Data can be Shared

Standards can be enforced

Security restrictions ca be applied

Integrity can be maintained

Conflicting requirements can be balanced

Data independence can be achieved.

DISADVANTAGES OF DBMS

A significant disadvantage of the DBMS system is cost. In addition to the cost of

purchasing of developing the software, the hardware has to be upgraded to allow for the

extensive programs and the workspace required for their execution and storage. While

centralization reduces duplication, the lack of duplication requires that the database be

adequately backed up so that in case of failure the data can be recovered.

FEATURES OF SQL SERVER (RDBMS)

SQL SERVER is one of the leading database management systems (DBMS) because it is

the only Database that meets the uncompromising requirements of today’s most demanding

information systems. From complex decision support systems (DSS) to the most rigorous online

transaction processing (OLTP) application, even application that require simultaneous DSS and

OLTP access to the same critical data, SQL Server leads the industry in both performance and

capability.

 

6. CONCLUSION

This project work is an attempt to develop a system that can be used for computerization of

activities in the steel company. Since these activities are tedious process requiring lot of effort,

more care has been taken for the system development. The requirements a suitable database is

created maximum effort were taken to avoid duplication in data entry and data storage. Various

reports can be generated by this system. The major advantage of the system is fast and accurate

in formation retrieval, minimization of clerical work, easy and efficient data storage and report

generation.

7. SCOPE FOR FUTURE ENHANCEMENTS

The “project title” in current developed using .NET, it can be future enhance to AJAX, THE

SYSTEM 2.0 or mobile applications for The system site development. Currently it is

implemented for single branch and it can be future enhanced for other branch too in future.

To our best knowledge the method what The system presented is new but still The system The

system using the classical tools of software engineering. The only modifications The system do

are using the super-characteristics for building the software quality model. By using this model

The system can bring good quality software. Model and the proposed solutions and opportunities

of improving software quality models.

A quality is an objective value dependent on sets of software attributes and customer’s

requirements. By applying the tools the quality of the software can be able to identify the

software type. And able to find the reusability of a software and understandability.

BIBILIOGRAPHY:

WEBSITES:

http://www.charlespetzold.com/dotnet http://www.csie.ntu.edu.tw/∼cjlin/libsvm/ http://msdn2.microsoft.com/en-us/vcsharp/aa336809.aspx https://minds.wisconsin.edu/bitstream/handle/. http://en.wikipedia.org/wiki/.NET_Framework https://www.aclweb.org/anthology http://support.microsoft.com/kb/318785 ijsetr.org/wp-content/uploads/.../IJSETR-VOL-4-ISSUE-4 http://msdn.microsoft.com/vstudio/express/visualcsharp http://msdn2.microsoft.com/library/aa388745.aspx

REFERENCE BOOK:

Microsoft ASP.NET and AJAX: Architecting Web Applications , by Dino Esposito SQL Server MVP Deep Dives , by Paul Nielson ASP.NET MVC Framework Unleashed , by Stephen Walther  C# 2008 for Programmers, 3rd Edition by Paul J. Deitel and Harvey M. Deitel

APPENDICES

A.SCREEN SHORT

B. SAMPLE CODING

using System;

usingSystem.Data;

usingSystem.Configuration;

usingSystem.Collections;

usingSystem.Web;

usingSystem.Web.Security;

usingSystem.Web.UI;

usingSystem.Web.UI.WebControls;

usingSystem.Web.UI.WebControls.WebParts;

usingSystem.Web.UI.HtmlControls;

usingSystem.Data.SqlClient;

publicpartialclassAdmin_a_transactiondetails : System.Web.UI.Page

{

SqlConnection con = newSqlConnection("Data Source=MAIN;InitialCatalog=attri;Integrated

Security=True");

protectedvoidPage_Load(object sender, EventArgs e)

{

display();

}

publicvoid display()

{

SqlDataAdapteradap = newSqlDataAdapter("select f.f_no,f.f_name,f.f_date,f.f_time,u.u_loginid

from a_fileuploadf,a_user u where f.u_code=u.u_code", con);

DataSet ds = newDataSet();

adap.Fill(ds);

GridView1.DataSource = ds;

GridView1.DataBind();

}

}

using System;

usingSystem.Data;

usingSystem.Configuration;

usingSystem.Collections;

usingSystem.Web;

usingSystem.Web.Security;

usingSystem.Web.UI;

usingSystem.Web.UI.WebControls;

usingSystem.Web.UI.WebControls.WebParts;

usingSystem.Web.UI.HtmlControls;

usingSystem.Data.SqlClient;

usingSystem.Windows.Forms;

publicpartialclassAdmin_a_userdetails : System.Web.UI.Page

{

SqlConnection con = newSqlConnection("Data Source=MAIN;InitialCatalog=attri;Integrated

Security=True");

protectedvoidPage_Load(object sender, EventArgs e)

{

if (!Page.IsPostBack)

{

display();

}

}

publicvoid display()

{

SqlDataAdapteradap = newSqlDataAdapter("select * from a_user", con);

DataSet ds = newDataSet();

adap.Fill(ds);

GridView1.DataSource = ds;

GridView1.DataBind();

}

protectedvoid GridView1_RowDeleting(object sender, GridViewDeleteEventArgs e)

{

if (MessageBox.Show("Do you want to Delete", "Delete Confirmation",

MessageBoxButtons.YesNo, MessageBoxIcon.Exclamation,

MessageBoxDefaultButton.Button2, MessageBoxOptions.DefaultDesktopOnly) ==

DialogResult.Yes)

{

GridViewRow row = GridView1.Rows[e.RowIndex];

string s = row.Cells[0].Text;

con.Open();

SqlCommandcmd = newSqlCommand("delete from a_user where u_code='" + s + "' ", con);

cmd.ExecuteNonQuery();

SqlCommand cmd1 = newSqlCommand("delete from a_key where u_code='" + s + "' ", con);

cmd1.ExecuteNonQuery();

con.Close();

}

display();

}

}

using System;

usingSystem.Data;

usingSystem.Configuration;

usingSystem.Collections;

usingSystem.Web;

usingSystem.Web.Security;

usingSystem.Web.UI;

usingSystem.Web.UI.WebControls;

usingSystem.Web.UI.WebControls.WebParts;

usingSystem.Web.UI.HtmlControls;

usingSystem.Data.SqlClient;

usingSystem.Security.Cryptography;

using System.Net;

using System.IO;

publicpartialclassAdmin_a_usercreate : System.Web.UI.Page

{

SqlConnection con = newSqlConnection("Data Source=MAIN;InitialCatalog=attri;Integrated

Security=True");

protectedvoidPage_Load(object sender, EventArgs e)

{

}

protectedvoidbtncreate_Click(object sender, EventArgs e)

{

SqlDataAdapterada = newSqlDataAdapter("select * from a_user where u_loginid='" + txtid.Text

+ "'", con);

DataSetdss = newDataSet();

ada.Fill(dss);

if (dss.Tables[0].Rows.Count == 0)

{

if (txtid.Text != ""&&txtpwd.Text != ""&&txtname.Text != ""&&txtaddress.Text !=

""&&txtcity.Text != ""&&txtpincode.Text != ""&&txtemailid.Text != ""&&txtcontactno.Text !=

"")

{

SqlDataAdapteradap = newSqlDataAdapter("select * from a_user where u_emailid='" +

txtemailid.Text + "'", con);

DataSet ds = newDataSet();

adap.Fill(ds);

if (ds.Tables[0].Rows.Count == 0)

{

con.Open();

SqlCommandcmd = newSqlCommand("insert into

a_user(u_loginid,u_pwd,u_name,u_address,u_city,u_pincode,u_emailid,u_contactno,u_status)

values('" + txtid.Text + "','" + txtpwd.Text + "','" + txtname.Text + "','" + txtaddress.Text + "','" +

txtcity.Text + "','" + txtpincode.Text + "','" + txtemailid.Text + "','" + txtcontactno.Text + "','" +

dd1.SelectedItem.ToString() + "')", con);

cmd.ExecuteNonQuery();

con.Close();

Upload("ftp://luisantsoftwares.com/PHR", "luisantcloud", "Luisant#123");

lblmsg.Visible = true;

lblmsg.Text = "Created Sucessfully";

txtpwd.Text = "";

txtname.Text = "";

txtaddress.Text = "";

txtcity.Text = "";

txtpincode.Text = "";

txtemailid.Text = "";

txtcontactno.Text = "";

}

else

{

lblmsg.Visible = true;

lblmsg.Text = "EmailD already exists";

}

}

else

{

lblmsg.Visible = true;

lblmsg.Text = "Enter all the values";

}

}

else

{

lblmsg.Visible = true;

lblmsg.Text = "Name already exists";

}

}

publicvoid Upload(stringurl, string user, stringpwd)

{

stringdir = txtid.Text;

FtpWebRequest request = (FtpWebRequest)WebRequest.Create(url + "/" + dir);

request.Method = WebRequestMethods.Ftp.MakeDirectory;

request.Credentials = newNetworkCredential(user, pwd);

FtpWebResponsemakeDirectoryResponse = (FtpWebResponse)request.GetResponse();

}

}